xref: /dpdk/lib/ptr_compress/rte_ptr_compress.h (revision fc10d6bd7632dff48a9f191e8d43872f47509e09)
1077596a4SPaul Szczepanek /* SPDX-License-Identifier: BSD-3-Clause
2077596a4SPaul Szczepanek  * Copyright(c) 2024 Arm Limited
3077596a4SPaul Szczepanek  */
4077596a4SPaul Szczepanek 
5077596a4SPaul Szczepanek #ifndef RTE_PTR_COMPRESS_H
6077596a4SPaul Szczepanek #define RTE_PTR_COMPRESS_H
7077596a4SPaul Szczepanek 
8077596a4SPaul Szczepanek /**
9077596a4SPaul Szczepanek  * @file
10077596a4SPaul Szczepanek  * Pointer compression and decompression functions.
11077596a4SPaul Szczepanek  *
12077596a4SPaul Szczepanek  * When passing arrays full of pointers between threads, memory containing
13077596a4SPaul Szczepanek  * the pointers is copied multiple times which is especially costly between
14077596a4SPaul Szczepanek  * cores. These functions allow us to compress the pointers.
15077596a4SPaul Szczepanek  *
16077596a4SPaul Szczepanek  * Compression takes advantage of the fact that pointers are usually located in
17077596a4SPaul Szczepanek  * a limited memory region. We compress them by converting them to offsets from
18077596a4SPaul Szczepanek  * a base memory address. Offsets can be stored in fewer bytes.
19077596a4SPaul Szczepanek  *
20077596a4SPaul Szczepanek  * The compression functions come in two varieties: 32-bit and 16-bit.
21077596a4SPaul Szczepanek  *
22077596a4SPaul Szczepanek  * To determine how many bits are needed to compress the pointer, calculate
23077596a4SPaul Szczepanek  * the biggest offset possible (highest value pointer - base pointer)
24077596a4SPaul Szczepanek  * and shift the value right according to alignment (shift by exponent of the
25077596a4SPaul Szczepanek  * power of 2 of alignment: aligned by 4 - shift by 2, aligned by 8 - shift by
26077596a4SPaul Szczepanek  * 3, etc.). The resulting value must fit in either 32 or 16 bits. You may
27077596a4SPaul Szczepanek  * use the macros provided in this file to do it programmatically.
28077596a4SPaul Szczepanek  *
29077596a4SPaul Szczepanek  * For usage example and further explanation please see this library's
30077596a4SPaul Szczepanek  * documentation in the programming guide.
31077596a4SPaul Szczepanek  */
32077596a4SPaul Szczepanek 
33077596a4SPaul Szczepanek #include <stdint.h>
34077596a4SPaul Szczepanek #include <inttypes.h>
35077596a4SPaul Szczepanek 
36077596a4SPaul Szczepanek #include <rte_bitops.h>
37077596a4SPaul Szczepanek #include <rte_branch_prediction.h>
38077596a4SPaul Szczepanek #include <rte_common.h>
39077596a4SPaul Szczepanek #include <rte_debug.h>
40077596a4SPaul Szczepanek #include <rte_vect.h>
41077596a4SPaul Szczepanek 
42077596a4SPaul Szczepanek #ifdef __cplusplus
43077596a4SPaul Szczepanek extern "C" {
44077596a4SPaul Szczepanek #endif
45077596a4SPaul Szczepanek 
46077596a4SPaul Szczepanek /**
47077596a4SPaul Szczepanek  * Calculate how many bits are required to store pointers within a given memory
48077596a4SPaul Szczepanek  * region as offsets. This can help decide which pointer compression functions
49077596a4SPaul Szczepanek  * can be used.
50077596a4SPaul Szczepanek  *
51077596a4SPaul Szczepanek  * @param mem_length
52077596a4SPaul Szczepanek  *   Length of the memory region the pointers are constrained to.
53077596a4SPaul Szczepanek  * @return
54077596a4SPaul Szczepanek  *   Number of bits required to store a value.
55077596a4SPaul Szczepanek  **/
56077596a4SPaul Szczepanek #define RTE_PTR_COMPRESS_BITS_NEEDED_FOR_POINTER_WITHIN_RANGE(mem_length) \
57077596a4SPaul Szczepanek 	(((uint64_t)mem_length) < 2 ? 1 : \
58077596a4SPaul Szczepanek 		(sizeof(uint64_t) * CHAR_BIT - \
59077596a4SPaul Szczepanek 		 rte_clz64((uint64_t)mem_length - 1)))
60077596a4SPaul Szczepanek 
61077596a4SPaul Szczepanek /**
62077596a4SPaul Szczepanek  * Calculate how many bits in the address can be dropped without losing any
63077596a4SPaul Szczepanek  * information thanks to the alignment of the address.
64077596a4SPaul Szczepanek  *
65077596a4SPaul Szczepanek  * @param alignment
66077596a4SPaul Szczepanek  *   Memory alignment.
67077596a4SPaul Szczepanek  * @return
68077596a4SPaul Szczepanek  *   Size of shift allowed without dropping any information from the pointer.
69077596a4SPaul Szczepanek  **/
70077596a4SPaul Szczepanek #define RTE_PTR_COMPRESS_BIT_SHIFT_FROM_ALIGNMENT(alignment) \
71077596a4SPaul Szczepanek 	((alignment) == 0 ? 0 : rte_ctz64((uint64_t)alignment))
72077596a4SPaul Szczepanek 
73077596a4SPaul Szczepanek /**
74077596a4SPaul Szczepanek  * Determine if rte_ptr_compress_16_shift can be used to compress pointers
75077596a4SPaul Szczepanek  * that contain addresses of memory objects whose memory is aligned by
76077596a4SPaul Szczepanek  * a given amount and contained in a given memory region.
77077596a4SPaul Szczepanek  *
78077596a4SPaul Szczepanek  * @param mem_length
79077596a4SPaul Szczepanek  *   The length of the memory region that contains the objects pointed to.
80077596a4SPaul Szczepanek  * @param obj_alignment
81077596a4SPaul Szczepanek  *   The alignment of objects pointed to.
82077596a4SPaul Szczepanek  * @return
83077596a4SPaul Szczepanek  *   1 if function can be used, 0 otherwise.
84077596a4SPaul Szczepanek  **/
85077596a4SPaul Szczepanek #define RTE_PTR_COMPRESS_CAN_COMPRESS_16_SHIFT(mem_length, obj_alignment) \
86077596a4SPaul Szczepanek 	((RTE_PTR_COMPRESS_BITS_NEEDED_FOR_POINTER_WITHIN_RANGE(mem_length) - \
87077596a4SPaul Szczepanek 	RTE_PTR_COMPRESS_BIT_SHIFT_FROM_ALIGNMENT(obj_alignment)) <= 16 ? 1 : 0)
88077596a4SPaul Szczepanek 
89077596a4SPaul Szczepanek /**
90077596a4SPaul Szczepanek  * Determine if rte_ptr_compress_32_shift can be used to compress pointers
91077596a4SPaul Szczepanek  * that contain addresses of memory objects whose memory is aligned by
92077596a4SPaul Szczepanek  * a given amount and contained in a given memory region.
93077596a4SPaul Szczepanek  *
94077596a4SPaul Szczepanek  * @param mem_length
95077596a4SPaul Szczepanek  *   The length of the memory region that contains the objects pointed to.
96077596a4SPaul Szczepanek  * @param obj_alignment
97077596a4SPaul Szczepanek  *   The alignment of objects pointed to.
98077596a4SPaul Szczepanek  * @return
99077596a4SPaul Szczepanek  *   1 if function can be used, 0 otherwise.
100077596a4SPaul Szczepanek  **/
101077596a4SPaul Szczepanek #define RTE_PTR_COMPRESS_CAN_COMPRESS_32_SHIFT(mem_length, obj_alignment) \
102077596a4SPaul Szczepanek 	((RTE_PTR_COMPRESS_BITS_NEEDED_FOR_POINTER_WITHIN_RANGE(mem_length) - \
103077596a4SPaul Szczepanek 	RTE_PTR_COMPRESS_BIT_SHIFT_FROM_ALIGNMENT(obj_alignment)) <= 32 ? 1 : 0)
104077596a4SPaul Szczepanek 
105077596a4SPaul Szczepanek /**
106077596a4SPaul Szczepanek  * Compress pointers into 32-bit offsets from base pointer.
107077596a4SPaul Szczepanek  *
108077596a4SPaul Szczepanek  * @note It is programmer's responsibility to ensure the resulting offsets fit
109077596a4SPaul Szczepanek  * into 32 bits. Alignment of the structures pointed to by the pointers allows
110077596a4SPaul Szczepanek  * us to drop bits from the offsets. This is controlled by the bit_shift
111077596a4SPaul Szczepanek  * parameter. This means that if structures are aligned by 8 bytes they must be
112077596a4SPaul Szczepanek  * within 32GB of the base pointer. If there is no such alignment guarantee they
113077596a4SPaul Szczepanek  * must be within 4GB.
114077596a4SPaul Szczepanek  *
115077596a4SPaul Szczepanek  * @param ptr_base
116077596a4SPaul Szczepanek  *   A pointer used to calculate offsets of pointers in src_table.
117077596a4SPaul Szczepanek  * @param src_table
118077596a4SPaul Szczepanek  *   A pointer to an array of pointers.
119077596a4SPaul Szczepanek  * @param dest_table
120077596a4SPaul Szczepanek  *   A pointer to an array of compressed pointers returned by this function.
121077596a4SPaul Szczepanek  * @param n
122077596a4SPaul Szczepanek  *   The number of objects to compress, must be strictly positive.
123077596a4SPaul Szczepanek  * @param bit_shift
124077596a4SPaul Szczepanek  *   Byte alignment of memory pointed to by the pointers allows for
125077596a4SPaul Szczepanek  *   bits to be dropped from the offset and hence widen the memory region that
126077596a4SPaul Szczepanek  *   can be covered. This controls how many bits are right shifted.
127077596a4SPaul Szczepanek  **/
128077596a4SPaul Szczepanek static __rte_always_inline void
rte_ptr_compress_32_shift(void * ptr_base,void * const * src_table,uint32_t * dest_table,size_t n,uint8_t bit_shift)129077596a4SPaul Szczepanek rte_ptr_compress_32_shift(void *ptr_base, void * const *src_table,
130077596a4SPaul Szczepanek 		uint32_t *dest_table, size_t n, uint8_t bit_shift)
131077596a4SPaul Szczepanek {
132077596a4SPaul Szczepanek 	size_t i = 0;
133077596a4SPaul Szczepanek #if defined RTE_HAS_SVE_ACLE && !defined RTE_ARCH_ARMv8_AARCH32
134077596a4SPaul Szczepanek 	svuint64_t v_ptr_table;
135077596a4SPaul Szczepanek 	do {
136077596a4SPaul Szczepanek 		svbool_t pg = svwhilelt_b64(i, n);
137077596a4SPaul Szczepanek 		v_ptr_table = svld1_u64(pg, (uint64_t *)src_table + i);
138077596a4SPaul Szczepanek 		v_ptr_table = svsub_x(pg, v_ptr_table, (uint64_t)ptr_base);
139077596a4SPaul Szczepanek 		v_ptr_table = svlsr_x(pg, v_ptr_table, bit_shift);
140077596a4SPaul Szczepanek 		svst1w(pg, &dest_table[i], v_ptr_table);
141077596a4SPaul Szczepanek 		i += svcntd();
142077596a4SPaul Szczepanek 	} while (i < n);
143077596a4SPaul Szczepanek #elif defined __ARM_NEON && !defined RTE_ARCH_ARMv8_AARCH32
144*fc10d6bdSPaul Szczepanek 	uintptr_t ptr_diff;
145077596a4SPaul Szczepanek 	uint64x2_t v_ptr_table;
146077596a4SPaul Szczepanek 	/* right shift is done by left shifting by negative int */
147077596a4SPaul Szczepanek 	int64x2_t v_shift = vdupq_n_s64(-bit_shift);
148077596a4SPaul Szczepanek 	uint64x2_t v_ptr_base = vdupq_n_u64((uint64_t)ptr_base);
149077596a4SPaul Szczepanek 	const size_t n_even = n & ~0x1;
150077596a4SPaul Szczepanek 	for (; i < n_even; i += 2) {
151077596a4SPaul Szczepanek 		v_ptr_table = vld1q_u64((const uint64_t *)src_table + i);
152077596a4SPaul Szczepanek 		v_ptr_table = vsubq_u64(v_ptr_table, v_ptr_base);
153077596a4SPaul Szczepanek 		v_ptr_table = vshlq_u64(v_ptr_table, v_shift);
154077596a4SPaul Szczepanek 		vst1_u32(dest_table + i, vqmovn_u64(v_ptr_table));
155077596a4SPaul Szczepanek 	}
156077596a4SPaul Szczepanek 	/* process leftover single item in case of odd number of n */
157077596a4SPaul Szczepanek 	if (unlikely(n & 0x1)) {
158077596a4SPaul Szczepanek 		ptr_diff = RTE_PTR_DIFF(src_table[i], ptr_base);
159077596a4SPaul Szczepanek 		dest_table[i] = (uint32_t) (ptr_diff >> bit_shift);
160077596a4SPaul Szczepanek 	}
161077596a4SPaul Szczepanek #else
162077596a4SPaul Szczepanek 	uintptr_t ptr_diff;
163077596a4SPaul Szczepanek 	for (; i < n; i++) {
164077596a4SPaul Szczepanek 		ptr_diff = RTE_PTR_DIFF(src_table[i], ptr_base);
165077596a4SPaul Szczepanek 		ptr_diff = ptr_diff >> bit_shift;
166077596a4SPaul Szczepanek 		RTE_ASSERT(ptr_diff <= UINT32_MAX);
167077596a4SPaul Szczepanek 		dest_table[i] = (uint32_t) ptr_diff;
168077596a4SPaul Szczepanek 	}
169077596a4SPaul Szczepanek #endif
170077596a4SPaul Szczepanek }
171077596a4SPaul Szczepanek 
172077596a4SPaul Szczepanek /**
173077596a4SPaul Szczepanek  * Decompress pointers from 32-bit offsets from base pointer.
174077596a4SPaul Szczepanek  *
175077596a4SPaul Szczepanek  * @param ptr_base
176077596a4SPaul Szczepanek  *   A pointer which was used to calculate offsets in src_table.
177077596a4SPaul Szczepanek  * @param src_table
178077596a4SPaul Szczepanek  *   A pointer to an array to compressed pointers.
179077596a4SPaul Szczepanek  * @param dest_table
180077596a4SPaul Szczepanek  *   A pointer to an array of decompressed pointers returned by this function.
181077596a4SPaul Szczepanek  * @param n
182077596a4SPaul Szczepanek  *   The number of objects to decompress, must be strictly positive.
183077596a4SPaul Szczepanek  * @param bit_shift
184077596a4SPaul Szczepanek  *   Byte alignment of memory pointed to by the pointers allows for
185077596a4SPaul Szczepanek  *   bits to be dropped from the offset and hence widen the memory region that
186077596a4SPaul Szczepanek  *   can be covered. This controls how many bits are left shifted when pointers
187077596a4SPaul Szczepanek  *   are recovered from the offsets.
188077596a4SPaul Szczepanek  **/
189077596a4SPaul Szczepanek static __rte_always_inline void
rte_ptr_decompress_32_shift(void * ptr_base,uint32_t const * src_table,void ** dest_table,size_t n,uint8_t bit_shift)190077596a4SPaul Szczepanek rte_ptr_decompress_32_shift(void *ptr_base, uint32_t const *src_table,
191077596a4SPaul Szczepanek 		void **dest_table, size_t n, uint8_t bit_shift)
192077596a4SPaul Szczepanek {
193077596a4SPaul Szczepanek 	size_t i = 0;
194077596a4SPaul Szczepanek #if defined RTE_HAS_SVE_ACLE && !defined RTE_ARCH_ARMv8_AARCH32
195077596a4SPaul Szczepanek 	svuint64_t v_ptr_table;
196077596a4SPaul Szczepanek 	do {
197077596a4SPaul Szczepanek 		svbool_t pg = svwhilelt_b64(i, n);
198077596a4SPaul Szczepanek 		v_ptr_table = svld1uw_u64(pg, &src_table[i]);
199077596a4SPaul Szczepanek 		v_ptr_table = svlsl_x(pg, v_ptr_table, bit_shift);
200077596a4SPaul Szczepanek 		v_ptr_table = svadd_x(pg, v_ptr_table, (uint64_t)ptr_base);
201077596a4SPaul Szczepanek 		svst1(pg, (uint64_t *)dest_table + i, v_ptr_table);
202077596a4SPaul Szczepanek 		i += svcntd();
203077596a4SPaul Szczepanek 	} while (i < n);
204077596a4SPaul Szczepanek #elif defined __ARM_NEON && !defined RTE_ARCH_ARMv8_AARCH32
205*fc10d6bdSPaul Szczepanek 	uintptr_t ptr_diff;
206077596a4SPaul Szczepanek 	uint64x2_t v_ptr_table;
207077596a4SPaul Szczepanek 	int64x2_t v_shift = vdupq_n_s64(bit_shift);
208077596a4SPaul Szczepanek 	uint64x2_t v_ptr_base = vdupq_n_u64((uint64_t)ptr_base);
209077596a4SPaul Szczepanek 	const size_t n_even = n & ~0x1;
210077596a4SPaul Szczepanek 	for (; i < n_even; i += 2) {
211077596a4SPaul Szczepanek 		v_ptr_table = vmovl_u32(vld1_u32(src_table + i));
212077596a4SPaul Szczepanek 		v_ptr_table = vshlq_u64(v_ptr_table, v_shift);
213077596a4SPaul Szczepanek 		v_ptr_table = vaddq_u64(v_ptr_table, v_ptr_base);
214077596a4SPaul Szczepanek 		vst1q_u64((uint64_t *)dest_table + i, v_ptr_table);
215077596a4SPaul Szczepanek 	}
216077596a4SPaul Szczepanek 	/* process leftover single item in case of odd number of n */
217077596a4SPaul Szczepanek 	if (unlikely(n & 0x1)) {
218*fc10d6bdSPaul Szczepanek 		ptr_diff = ((uintptr_t) src_table[i]) << bit_shift;
219077596a4SPaul Szczepanek 		dest_table[i] = RTE_PTR_ADD(ptr_base, ptr_diff);
220077596a4SPaul Szczepanek 	}
221077596a4SPaul Szczepanek #else
222077596a4SPaul Szczepanek 	uintptr_t ptr_diff;
223077596a4SPaul Szczepanek 	for (; i < n; i++) {
224077596a4SPaul Szczepanek 		ptr_diff = ((uintptr_t) src_table[i]) << bit_shift;
225077596a4SPaul Szczepanek 		dest_table[i] = RTE_PTR_ADD(ptr_base, ptr_diff);
226077596a4SPaul Szczepanek 	}
227077596a4SPaul Szczepanek #endif
228077596a4SPaul Szczepanek }
229077596a4SPaul Szczepanek 
230077596a4SPaul Szczepanek /**
231077596a4SPaul Szczepanek  * Compress pointers into 16-bit offsets from base pointer.
232077596a4SPaul Szczepanek  *
233077596a4SPaul Szczepanek  * @note It is programmer's responsibility to ensure the resulting offsets fit
234077596a4SPaul Szczepanek  * into 16 bits. Alignment of the structures pointed to by the pointers allows
235077596a4SPaul Szczepanek  * us to drop bits from the offsets. This is controlled by the bit_shift
236077596a4SPaul Szczepanek  * parameter. This means that if structures are aligned by 8 bytes they must be
237077596a4SPaul Szczepanek  * within 256KB of the base pointer. If there is no such alignment guarantee
238077596a4SPaul Szczepanek  * they must be within 64KB.
239077596a4SPaul Szczepanek  *
240077596a4SPaul Szczepanek  * @param ptr_base
241077596a4SPaul Szczepanek  *   A pointer used to calculate offsets of pointers in src_table.
242077596a4SPaul Szczepanek  * @param src_table
243077596a4SPaul Szczepanek  *   A pointer to an array of pointers.
244077596a4SPaul Szczepanek  * @param dest_table
245077596a4SPaul Szczepanek  *   A pointer to an array of compressed pointers returned by this function.
246077596a4SPaul Szczepanek  * @param n
247077596a4SPaul Szczepanek  *   The number of objects to compress, must be strictly positive.
248077596a4SPaul Szczepanek  * @param bit_shift
249077596a4SPaul Szczepanek  *   Byte alignment of memory pointed to by the pointers allows for
250077596a4SPaul Szczepanek  *   bits to be dropped from the offset and hence widen the memory region that
251077596a4SPaul Szczepanek  *   can be covered. This controls how many bits are right shifted.
252077596a4SPaul Szczepanek  **/
253077596a4SPaul Szczepanek static __rte_always_inline void
rte_ptr_compress_16_shift(void * ptr_base,void * const * src_table,uint16_t * dest_table,size_t n,uint8_t bit_shift)254077596a4SPaul Szczepanek rte_ptr_compress_16_shift(void *ptr_base, void * const *src_table,
255077596a4SPaul Szczepanek 		uint16_t *dest_table, size_t n, uint8_t bit_shift)
256077596a4SPaul Szczepanek {
257077596a4SPaul Szczepanek 
258077596a4SPaul Szczepanek 	size_t i = 0;
259077596a4SPaul Szczepanek #if defined RTE_HAS_SVE_ACLE && !defined RTE_ARCH_ARMv8_AARCH32
260077596a4SPaul Szczepanek 	svuint64_t v_ptr_table;
261077596a4SPaul Szczepanek 	do {
262077596a4SPaul Szczepanek 		svbool_t pg = svwhilelt_b64(i, n);
263077596a4SPaul Szczepanek 		v_ptr_table = svld1_u64(pg, (uint64_t *)src_table + i);
264077596a4SPaul Szczepanek 		v_ptr_table = svsub_x(pg, v_ptr_table, (uint64_t)ptr_base);
265077596a4SPaul Szczepanek 		v_ptr_table = svlsr_x(pg, v_ptr_table, bit_shift);
266077596a4SPaul Szczepanek 		svst1h(pg, &dest_table[i], v_ptr_table);
267077596a4SPaul Szczepanek 		i += svcntd();
268077596a4SPaul Szczepanek 	} while (i < n);
269077596a4SPaul Szczepanek #else
270077596a4SPaul Szczepanek 	uintptr_t ptr_diff;
271077596a4SPaul Szczepanek 	for (; i < n; i++) {
272077596a4SPaul Szczepanek 		ptr_diff = RTE_PTR_DIFF(src_table[i], ptr_base);
273077596a4SPaul Szczepanek 		ptr_diff = ptr_diff >> bit_shift;
274077596a4SPaul Szczepanek 		RTE_ASSERT(ptr_diff <= UINT16_MAX);
275077596a4SPaul Szczepanek 		dest_table[i] = (uint16_t) ptr_diff;
276077596a4SPaul Szczepanek 	}
277077596a4SPaul Szczepanek #endif
278077596a4SPaul Szczepanek }
279077596a4SPaul Szczepanek 
280077596a4SPaul Szczepanek /**
281077596a4SPaul Szczepanek  * Decompress pointers from 16-bit offsets from base pointer.
282077596a4SPaul Szczepanek  *
283077596a4SPaul Szczepanek  * @param ptr_base
284077596a4SPaul Szczepanek  *   A pointer which was used to calculate offsets in src_table.
285077596a4SPaul Szczepanek  * @param src_table
286077596a4SPaul Szczepanek  *   A pointer to an array to compressed pointers.
287077596a4SPaul Szczepanek  * @param dest_table
288077596a4SPaul Szczepanek  *   A pointer to an array of decompressed pointers returned by this function.
289077596a4SPaul Szczepanek  * @param n
290077596a4SPaul Szczepanek  *   The number of objects to decompress, must be strictly positive.
291077596a4SPaul Szczepanek  * @param bit_shift
292077596a4SPaul Szczepanek  *   Byte alignment of memory pointed to by the pointers allows for
293077596a4SPaul Szczepanek  *   bits to be dropped from the offset and hence widen the memory region that
294077596a4SPaul Szczepanek  *   can be covered. This controls how many bits are left shifted when pointers
295077596a4SPaul Szczepanek  *   are recovered from the offsets.
296077596a4SPaul Szczepanek  **/
297077596a4SPaul Szczepanek static __rte_always_inline void
rte_ptr_decompress_16_shift(void * ptr_base,uint16_t const * src_table,void ** dest_table,size_t n,uint8_t bit_shift)298077596a4SPaul Szczepanek rte_ptr_decompress_16_shift(void *ptr_base, uint16_t const *src_table,
299077596a4SPaul Szczepanek 		void **dest_table, size_t n, uint8_t bit_shift)
300077596a4SPaul Szczepanek {
301077596a4SPaul Szczepanek 	size_t i = 0;
302077596a4SPaul Szczepanek #if defined RTE_HAS_SVE_ACLE && !defined RTE_ARCH_ARMv8_AARCH32
303077596a4SPaul Szczepanek 	svuint64_t v_ptr_table;
304077596a4SPaul Szczepanek 	do {
305077596a4SPaul Szczepanek 		svbool_t pg = svwhilelt_b64(i, n);
306077596a4SPaul Szczepanek 		v_ptr_table = svld1uh_u64(pg, &src_table[i]);
307077596a4SPaul Szczepanek 		v_ptr_table = svlsl_x(pg, v_ptr_table, bit_shift);
308077596a4SPaul Szczepanek 		v_ptr_table = svadd_x(pg, v_ptr_table, (uint64_t)ptr_base);
309077596a4SPaul Szczepanek 		svst1(pg, (uint64_t *)dest_table + i, v_ptr_table);
310077596a4SPaul Szczepanek 		i += svcntd();
311077596a4SPaul Szczepanek 	} while (i < n);
312077596a4SPaul Szczepanek #else
313077596a4SPaul Szczepanek 	uintptr_t ptr_diff;
314077596a4SPaul Szczepanek 	for (; i < n; i++) {
315077596a4SPaul Szczepanek 		ptr_diff = ((uintptr_t) src_table[i]) << bit_shift;
316077596a4SPaul Szczepanek 		dest_table[i] = RTE_PTR_ADD(ptr_base, ptr_diff);
317077596a4SPaul Szczepanek 	}
318077596a4SPaul Szczepanek #endif
319077596a4SPaul Szczepanek }
320077596a4SPaul Szczepanek 
321077596a4SPaul Szczepanek #ifdef __cplusplus
322077596a4SPaul Szczepanek }
323077596a4SPaul Szczepanek #endif
324077596a4SPaul Szczepanek 
325077596a4SPaul Szczepanek #endif /* RTE_PTR_COMPRESS_H */
326