@@ -6137,7 +6137,7 @@ SZ_PUBLIC void sz_fill_sve(sz_ptr_t target, sz_size_t length, sz_u8_t value) {
6137
6137
6138
6138
if (length <= vec_len) {
6139
6139
// Small buffer case: use mask to handle small writes
6140
- svbool_t mask = svwhilelt_b8 ((sz_size_t )0ull , length);
6140
+ svbool_t mask = svwhilelt_b8 ((sz_u64_t )0ull , ( sz_u64_t ) length);
6141
6141
svst1_u8 (mask, (unsigned char *)target, value_vec);
6142
6142
}
6143
6143
else {
@@ -6147,7 +6147,7 @@ SZ_PUBLIC void sz_fill_sve(sz_ptr_t target, sz_size_t length, sz_u8_t value) {
6147
6147
sz_size_t body_length = length - head_length - tail_length;
6148
6148
6149
6149
// Handle unaligned head
6150
- svbool_t head_mask = svwhilelt_b8 ((sz_size_t )0ull , head_length);
6150
+ svbool_t head_mask = svwhilelt_b8 ((sz_u64_t )0ull , ( sz_u64_t ) head_length);
6151
6151
svst1_u8 (head_mask, (unsigned char *)target, value_vec);
6152
6152
target += head_length;
6153
6153
@@ -6157,7 +6157,7 @@ SZ_PUBLIC void sz_fill_sve(sz_ptr_t target, sz_size_t length, sz_u8_t value) {
6157
6157
}
6158
6158
6159
6159
// Handle unaligned tail
6160
- svbool_t tail_mask = svwhilelt_b8 ((sz_size_t )0ull , tail_length);
6160
+ svbool_t tail_mask = svwhilelt_b8 ((sz_u64_t )0ull , ( sz_u64_t ) tail_length);
6161
6161
svst1_u8 (tail_mask, (unsigned char *)target, value_vec);
6162
6162
}
6163
6163
}
@@ -6174,7 +6174,7 @@ SZ_PUBLIC void sz_copy_sve(sz_ptr_t target, sz_cptr_t source, sz_size_t length)
6174
6174
// When the buffer is small, there isn't much to innovate.
6175
6175
if (length <= vec_len) {
6176
6176
// Small buffer case: use mask to handle small writes
6177
- svbool_t mask = svwhilelt_b8 ((sz_size_t )0ull , length);
6177
+ svbool_t mask = svwhilelt_b8 ((sz_u64_t )0ull , ( sz_u64_t ) length);
6178
6178
svuint8_t data = svld1_u8 (mask, (unsigned char *)source);
6179
6179
svst1_u8 (mask, (unsigned char *)target, data);
6180
6180
}
@@ -6199,10 +6199,10 @@ SZ_PUBLIC void sz_copy_sve(sz_ptr_t target, sz_cptr_t source, sz_size_t length)
6199
6199
sz_size_t body_length = length - head_length - tail_length;
6200
6200
6201
6201
// Handle unaligned parts
6202
- svbool_t head_mask = svwhilelt_b8 ((sz_size_t )0ull , head_length);
6202
+ svbool_t head_mask = svwhilelt_b8 ((sz_u64_t )0ull , ( sz_u64_t ) head_length);
6203
6203
svuint8_t head_data = svld1_u8 (head_mask, (unsigned char *)source);
6204
6204
svst1_u8 (head_mask, (unsigned char *)target, head_data);
6205
- svbool_t tail_mask = svwhilelt_b8 ((sz_size_t )0ull , tail_length);
6205
+ svbool_t tail_mask = svwhilelt_b8 ((sz_u64_t )0ull , ( sz_u64_t ) tail_length);
6206
6206
svuint8_t tail_data = svld1_u8 (tail_mask, (unsigned char *)source + head_length + body_length);
6207
6207
svst1_u8 (tail_mask, (unsigned char *)target + head_length + body_length, tail_data);
6208
6208
target += head_length;
@@ -6218,15 +6218,15 @@ SZ_PUBLIC void sz_copy_sve(sz_ptr_t target, sz_cptr_t source, sz_size_t length)
6218
6218
// Up to (vec_len * 2 - 1) bytes of data may be left in the body,
6219
6219
// so we can unroll the last two optional loop iterations.
6220
6220
if (body_length > vec_len) {
6221
- svbool_t mask = svwhilelt_b8 ((sz_size_t )0ull , body_length);
6221
+ svbool_t mask = svwhilelt_b8 ((sz_u64_t )0ull , ( sz_u64_t ) body_length);
6222
6222
svuint8_t data = svld1_u8 (mask, (unsigned char *)source);
6223
6223
svst1_u8 (mask, (unsigned char *)target, data);
6224
6224
body_length -= vec_len;
6225
6225
source += body_length;
6226
6226
target += body_length;
6227
6227
}
6228
6228
if (body_length) {
6229
- svbool_t mask = svwhilelt_b8 ((sz_size_t )0ull , body_length);
6229
+ svbool_t mask = svwhilelt_b8 ((sz_u64_t )0ull , ( sz_u64_t ) body_length);
6230
6230
svuint8_t data = svld1_u8 (mask, (unsigned char *)source);
6231
6231
svst1_u8 (mask, (unsigned char *)target, data);
6232
6232
}
0 commit comments