numam-spdk/lib/util/base64_sve.c
Rui Chang d483d8a4f1 lib/util: optimize base64 encode and decode using ARM SVE intrinsics
Optimize base64 encode and decode using ARM SVE intrinsics.

Signed-off-by: Rui Chang <rui.chang@arm.com>
Change-Id: I63ae5b9105aef690e1fad8ec1ef98c8339758e9f
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/8858
Community-CI: Broadcom CI <spdk-ci.pdl@broadcom.com>
Community-CI: Mellanox Build Bot
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Reviewed-by: Ziye Yang <ziye.yang@intel.com>
Reviewed-by: Paul Luse <paul.e.luse@intel.com>
Reviewed-by: Aleksey Marchuk <alexeymar@mellanox.com>
Reviewed-by: Jim Harris <james.r.harris@intel.com>
2021-08-06 07:31:06 +00:00

501 lines
17 KiB
C

/*-
* BSD LICENSE
*
* Copyright(c) ARM Limited. 2021 All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are
* met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS
* IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A
* PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
* TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
* PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
* NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
* SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef __aarch64__
#error Unsupported hardware
#endif
#include "spdk/stdinc.h"
#include <arm_sve.h>
static int
table_lookup_8vec(svuint8_t tbl_vec0, svuint8_t tbl_vec1, svuint8_t tbl_vec2, svuint8_t tbl_vec3,
svuint8_t tbl_vec4, svuint8_t tbl_vec5, svuint8_t tbl_vec6, svuint8_t tbl_vec7,
svuint8_t indices, svuint8_t *output, svbool_t p8_in, uint64_t vl)
{
svuint8_t res2, res3, res4, res5, res6, res7;
/*
* In base64 decode table, the first 32 elements are invalid value,
* so skip tbl_vec0 and tbl_vec1
*/
indices = svsub_n_u8_z(p8_in, indices, 2 * vl);
res2 = svtbl_u8(tbl_vec2, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res3 = svtbl_u8(tbl_vec3, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res4 = svtbl_u8(tbl_vec4, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res5 = svtbl_u8(tbl_vec5, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res6 = svtbl_u8(tbl_vec6, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res7 = svtbl_u8(tbl_vec7, indices);
*output = svdup_n_u8(0);
*output = svadd_u8_z(p8_in, res2, *output);
*output = svadd_u8_z(p8_in, res3, *output);
*output = svadd_u8_z(p8_in, res4, *output);
*output = svadd_u8_z(p8_in, res5, *output);
*output = svadd_u8_z(p8_in, res6, *output);
*output = svadd_u8_z(p8_in, res7, *output);
if (svcntp_b8(p8_in, svcmpeq_n_u8(p8_in, *output, 255))) {
return -1;
}
return 0;
}
static int
table_lookup_4vec(svuint8_t tbl_vec0, svuint8_t tbl_vec1, svuint8_t tbl_vec2, svuint8_t tbl_vec3,
svuint8_t indices, svuint8_t *output, svbool_t p8_in, uint64_t vl)
{
svuint8_t res0, res1, res2, res3;
res0 = svtbl_u8(tbl_vec0, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res1 = svtbl_u8(tbl_vec1, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res2 = svtbl_u8(tbl_vec2, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res3 = svtbl_u8(tbl_vec3, indices);
*output = svdup_n_u8(0);
*output = svadd_u8_z(p8_in, res0, *output);
*output = svadd_u8_z(p8_in, res1, *output);
*output = svadd_u8_z(p8_in, res2, *output);
*output = svadd_u8_z(p8_in, res3, *output);
if (svcntp_b8(p8_in, svcmpeq_n_u8(p8_in, *output, 255))) {
return -1;
}
return 0;
}
static int
table_lookup_3vec(svuint8_t tbl_vec0, svuint8_t tbl_vec1, svuint8_t tbl_vec2, svuint8_t indices,
svuint8_t *output, svbool_t p8_in, uint64_t vl)
{
svuint8_t res0, res1, res2;
res0 = svtbl_u8(tbl_vec0, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res1 = svtbl_u8(tbl_vec1, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res2 = svtbl_u8(tbl_vec2, indices);
*output = svdup_n_u8(0);
*output = svadd_u8_z(p8_in, res0, *output);
*output = svadd_u8_z(p8_in, res1, *output);
*output = svadd_u8_z(p8_in, res2, *output);
if (svcntp_b8(p8_in, svcmpeq_n_u8(p8_in, *output, 255))) {
return -1;
}
return 0;
}
static int
table_lookup_2vec(svuint8_t tbl_vec0, svuint8_t tbl_vec1, svuint8_t indices, svuint8_t *output,
svbool_t p8_in, uint64_t vl)
{
svuint8_t res0, res1;
res0 = svtbl_u8(tbl_vec0, indices);
indices = svsub_n_u8_z(p8_in, indices, vl);
res1 = svtbl_u8(tbl_vec1, indices);
*output = svdup_n_u8(0);
*output = svadd_u8_z(p8_in, res0, *output);
*output = svadd_u8_z(p8_in, res1, *output);
if (svcntp_b8(p8_in, svcmpeq_n_u8(p8_in, *output, 255))) {
return -1;
}
return 0;
}
static inline void
convert_6bits_to_8bits(svbool_t pred, uint8_t *src, svuint8_t *temp0, svuint8_t *temp1,
svuint8_t *temp2, svuint8_t *temp3)
{
svuint8_t str0, str1, str2;
svuint8x3_t ld_enc_input;
ld_enc_input = svld3_u8(pred, src);
str0 = svget3_u8(ld_enc_input, 0);
str1 = svget3_u8(ld_enc_input, 1);
str2 = svget3_u8(ld_enc_input, 2);
*temp0 = svlsr_n_u8_z(pred, str0, 2);
*temp1 = svand_u8_z(pred, svorr_u8_z(pred, svlsr_n_u8_z(pred, str1, 4), svlsl_n_u8_z(pred, str0,
4)),
svdup_u8(0x3F));
*temp2 = svand_u8_z(pred, svorr_u8_z(pred, svlsr_n_u8_z(pred, str2, 6), svlsl_n_u8_z(pred, str1,
2)),
svdup_u8(0x3F));
*temp3 = svand_u8_z(pred, str2, svdup_u8(0x3F));
}
static inline void
convert_8bits_to_6bits(svbool_t pred, svuint8_t temp0, svuint8_t temp1, svuint8_t temp2,
svuint8_t temp3, svuint8_t *output0, svuint8_t *output1, svuint8_t *output2)
{
*output0 = svorr_u8_z(pred, svlsl_n_u8_z(pred, temp0, 2), svlsr_n_u8_z(pred, temp1, 4));
*output1 = svorr_u8_z(pred, svlsl_n_u8_z(pred, temp1, 4), svlsr_n_u8_z(pred, temp2, 2));
*output2 = svorr_u8_z(pred, svlsl_n_u8_z(pred, temp2, 6), temp3);
}
static void
base64_encode_sve(char **dst, const char *enc_table, const void **src, size_t *src_len)
{
uint64_t vl = svcntb();
svuint8_t temp0, temp1, temp2, temp3;
svuint8_t output0, output1, output2, output3;
svuint8_t tbl_enc0, tbl_enc1, tbl_enc2, tbl_enc3;
svuint8x4_t st_enc_output;
svbool_t p8_all = svptrue_b8();
svbool_t pred;
uint64_t i = 0;
uint64_t pred_count = 0;
uint64_t N = (*src_len / 3) * 3;
if (vl == 16) {
tbl_enc0 = svld1_u8(p8_all, (uint8_t *)enc_table + 0);
tbl_enc1 = svld1_u8(p8_all, (uint8_t *)enc_table + 16);
tbl_enc2 = svld1_u8(p8_all, (uint8_t *)enc_table + 32);
tbl_enc3 = svld1_u8(p8_all, (uint8_t *)enc_table + 48);
while (i < N) {
pred = svwhilelt_b8(i / 3, N / 3);
convert_6bits_to_8bits(pred, (uint8_t *)*src, &temp0, &temp1, &temp2, &temp3);
table_lookup_4vec(tbl_enc0, tbl_enc1, tbl_enc2, tbl_enc3, temp0, &output0, pred, vl);
table_lookup_4vec(tbl_enc0, tbl_enc1, tbl_enc2, tbl_enc3, temp1, &output1, pred, vl);
table_lookup_4vec(tbl_enc0, tbl_enc1, tbl_enc2, tbl_enc3, temp2, &output2, pred, vl);
table_lookup_4vec(tbl_enc0, tbl_enc1, tbl_enc2, tbl_enc3, temp3, &output3, pred, vl);
st_enc_output = svcreate4_u8(output0, output1, output2, output3);
svst4_u8(pred, (uint8_t *)*dst, st_enc_output);
pred_count = svcntp_b8(pred, pred);
*src += pred_count * 3;
*dst += pred_count * 4;
*src_len -= pred_count * 3;
i += pred_count * 3;
}
} else if (vl == 32 || vl == 48) {
tbl_enc0 = svld1_u8(p8_all, (uint8_t *)enc_table + 0);
pred = svwhilelt_b8(vl, (uint64_t)64);
tbl_enc1 = svld1_u8(pred, (uint8_t *)enc_table + vl);
while (i < N) {
pred = svwhilelt_b8(i / 3, N / 3);
convert_6bits_to_8bits(pred, (uint8_t *)*src, &temp0, &temp1, &temp2, &temp3);
table_lookup_2vec(tbl_enc0, tbl_enc1, temp0, &output0, pred, vl);
table_lookup_2vec(tbl_enc0, tbl_enc1, temp1, &output1, pred, vl);
table_lookup_2vec(tbl_enc0, tbl_enc1, temp2, &output2, pred, vl);
table_lookup_2vec(tbl_enc0, tbl_enc1, temp3, &output3, pred, vl);
st_enc_output = svcreate4_u8(output0, output1, output2, output3);
svst4_u8(pred, (uint8_t *)*dst, st_enc_output);
pred_count = svcntp_b8(pred, pred);
*src += pred_count * 3;
*dst += pred_count * 4;
*src_len -= pred_count * 3;
i += pred_count * 3;
}
} else if (vl >= 64) {
pred = svwhilelt_b8((uint64_t)0, (uint64_t)64);
tbl_enc0 = svld1_u8(pred, (uint8_t *)enc_table);
while (i < N) {
pred = svwhilelt_b8(i / 3, N / 3);
convert_6bits_to_8bits(pred, (uint8_t *)*src, &temp0, &temp1, &temp2, &temp3);
output0 = svtbl_u8(tbl_enc0, temp0);
output1 = svtbl_u8(tbl_enc0, temp1);
output2 = svtbl_u8(tbl_enc0, temp2);
output3 = svtbl_u8(tbl_enc0, temp3);
st_enc_output = svcreate4_u8(output0, output1, output2, output3);
svst4_u8(pred, (uint8_t *)*dst, st_enc_output);
pred_count = svcntp_b8(pred, pred);
*src += pred_count * 3;
*dst += pred_count * 4;
*src_len -= pred_count * 3;
i += pred_count * 3;
}
}
}
static void
base64_decode_sve(void **dst, const uint8_t *dec_table, const uint8_t **src, size_t *src_len)
{
uint64_t vl = svcntb();
svuint8_t str0, str1, str2, str3;
svuint8_t temp0, temp1, temp2, temp3;
svuint8_t output0, output1, output2;
svuint8_t tbl_dec0, tbl_dec1, tbl_dec2, tbl_dec3, tbl_dec4, tbl_dec5, tbl_dec6, tbl_dec7;
svuint8x3_t st_dec_output;
svbool_t p8_all = svptrue_b8();
svbool_t pred;
uint64_t i = 0;
uint64_t pred_count = 0;
uint64_t N = (*src_len / 4) * 4;
svuint8x4_t ld_dec_input;
if (vl == 16) {
tbl_dec0 = svld1_u8(p8_all, (uint8_t *)dec_table + 0);
tbl_dec1 = svld1_u8(p8_all, (uint8_t *)dec_table + 16);
tbl_dec2 = svld1_u8(p8_all, (uint8_t *)dec_table + 32);
tbl_dec3 = svld1_u8(p8_all, (uint8_t *)dec_table + 48);
tbl_dec4 = svld1_u8(p8_all, (uint8_t *)dec_table + 64);
tbl_dec5 = svld1_u8(p8_all, (uint8_t *)dec_table + 80);
tbl_dec6 = svld1_u8(p8_all, (uint8_t *)dec_table + 96);
tbl_dec7 = svld1_u8(p8_all, (uint8_t *)dec_table + 112);
while (i < N) {
pred = svwhilelt_b8(i / 4, N / 4);
ld_dec_input = svld4_u8(pred, *src);
str0 = svget4_u8(ld_dec_input, 0);
str1 = svget4_u8(ld_dec_input, 1);
str2 = svget4_u8(ld_dec_input, 2);
str3 = svget4_u8(ld_dec_input, 3);
if (svcntp_b8(pred, svcmpge_n_u8(pred, str0, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str1, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str2, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str3, 128))) { return; }
if (table_lookup_8vec(tbl_dec0, tbl_dec1, tbl_dec2, tbl_dec3, tbl_dec4, tbl_dec5, tbl_dec6,
tbl_dec7, str0, &temp0, pred, vl)) { return; }
if (table_lookup_8vec(tbl_dec0, tbl_dec1, tbl_dec2, tbl_dec3, tbl_dec4, tbl_dec5, tbl_dec6,
tbl_dec7, str1, &temp1, pred, vl)) { return; }
if (table_lookup_8vec(tbl_dec0, tbl_dec1, tbl_dec2, tbl_dec3, tbl_dec4, tbl_dec5, tbl_dec6,
tbl_dec7, str2, &temp2, pred, vl)) { return; }
if (table_lookup_8vec(tbl_dec0, tbl_dec1, tbl_dec2, tbl_dec3, tbl_dec4, tbl_dec5, tbl_dec6,
tbl_dec7, str3, &temp3, pred, vl)) { return; }
convert_8bits_to_6bits(pred, temp0, temp1, temp2, temp3, &output0, &output1, &output2);
st_dec_output = svcreate3_u8(output0, output1, output2);
svst3_u8(pred, (uint8_t *)*dst, st_dec_output);
pred_count = svcntp_b8(pred, pred);
*src += pred_count * 4;
*dst += pred_count * 3;
*src_len -= pred_count * 4;
i += pred_count * 4;
}
} else if (vl == 32) {
tbl_dec0 = svld1_u8(p8_all, (uint8_t *)dec_table + 0);
tbl_dec1 = svld1_u8(p8_all, (uint8_t *)dec_table + vl);
tbl_dec2 = svld1_u8(p8_all, (uint8_t *)dec_table + vl * 2);
tbl_dec3 = svld1_u8(p8_all, (uint8_t *)dec_table + vl * 3);
while (i < N) {
pred = svwhilelt_b8(i / 4, N / 4);
ld_dec_input = svld4_u8(pred, *src);
str0 = svget4_u8(ld_dec_input, 0);
str1 = svget4_u8(ld_dec_input, 1);
str2 = svget4_u8(ld_dec_input, 2);
str3 = svget4_u8(ld_dec_input, 3);
if (svcntp_b8(pred, svcmpge_n_u8(pred, str0, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str1, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str2, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str3, 128))) { return; }
if (table_lookup_4vec(tbl_dec0, tbl_dec1, tbl_dec2, tbl_dec3, str0, &temp0, pred, vl)) { return; }
if (table_lookup_4vec(tbl_dec0, tbl_dec1, tbl_dec2, tbl_dec3, str1, &temp1, pred, vl)) { return; }
if (table_lookup_4vec(tbl_dec0, tbl_dec1, tbl_dec2, tbl_dec3, str2, &temp2, pred, vl)) { return; }
if (table_lookup_4vec(tbl_dec0, tbl_dec1, tbl_dec2, tbl_dec3, str3, &temp3, pred, vl)) { return; }
convert_8bits_to_6bits(pred, temp0, temp1, temp2, temp3, &output0, &output1, &output2);
st_dec_output = svcreate3_u8(output0, output1, output2);
svst3_u8(pred, (uint8_t *)*dst, st_dec_output);
pred_count = svcntp_b8(pred, pred);
*src += pred_count * 4;
*dst += pred_count * 3;
*src_len -= pred_count * 4;
i += pred_count * 4;
}
} else if (vl == 48) {
tbl_dec0 = svld1_u8(p8_all, (uint8_t *)dec_table + 0);
tbl_dec1 = svld1_u8(p8_all, (uint8_t *)dec_table + vl);
pred = svwhilelt_b8(vl * 2, (uint64_t)128);
tbl_dec2 = svld1_u8(pred, (uint8_t *)dec_table + 2 * vl);
while (i < N) {
pred = svwhilelt_b8(i / 4, N / 4);
ld_dec_input = svld4_u8(pred, *src);
str0 = svget4_u8(ld_dec_input, 0);
str1 = svget4_u8(ld_dec_input, 1);
str2 = svget4_u8(ld_dec_input, 2);
str3 = svget4_u8(ld_dec_input, 3);
if (svcntp_b8(pred, svcmpge_n_u8(pred, str0, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str1, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str2, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str3, 128))) { return; }
if (table_lookup_3vec(tbl_dec0, tbl_dec1, tbl_dec2, str0, &temp0, pred, vl)) { return; }
if (table_lookup_3vec(tbl_dec0, tbl_dec1, tbl_dec2, str1, &temp1, pred, vl)) { return; }
if (table_lookup_3vec(tbl_dec0, tbl_dec1, tbl_dec2, str2, &temp2, pred, vl)) { return; }
if (table_lookup_3vec(tbl_dec0, tbl_dec1, tbl_dec2, str3, &temp3, pred, vl)) { return; }
convert_8bits_to_6bits(pred, temp0, temp1, temp2, temp3, &output0, &output1, &output2);
st_dec_output = svcreate3_u8(output0, output1, output2);
svst3_u8(pred, (uint8_t *)*dst, st_dec_output);
pred_count = svcntp_b8(pred, pred);
*src += pred_count * 4;
*dst += pred_count * 3;
*src_len -= pred_count * 4;
i += pred_count * 4;
}
} else if (vl == 64 || vl == 80 || vl == 96 || vl == 112) {
tbl_dec0 = svld1_u8(p8_all, (uint8_t *)dec_table + 0);
pred = svwhilelt_b8(vl, (uint64_t)128);
tbl_dec1 = svld1_u8(pred, (uint8_t *)dec_table + vl);
while (i < N) {
pred = svwhilelt_b8(i / 4, N / 4);
ld_dec_input = svld4_u8(pred, *src);
str0 = svget4_u8(ld_dec_input, 0);
str1 = svget4_u8(ld_dec_input, 1);
str2 = svget4_u8(ld_dec_input, 2);
str3 = svget4_u8(ld_dec_input, 3);
if (svcntp_b8(pred, svcmpge_n_u8(pred, str0, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str1, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str2, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str3, 128))) { return; }
if (table_lookup_2vec(tbl_dec0, tbl_dec1, str0, &temp0, pred, vl)) { return; }
if (table_lookup_2vec(tbl_dec0, tbl_dec1, str1, &temp1, pred, vl)) { return; }
if (table_lookup_2vec(tbl_dec0, tbl_dec1, str2, &temp2, pred, vl)) { return; }
if (table_lookup_2vec(tbl_dec0, tbl_dec1, str3, &temp3, pred, vl)) { return; }
convert_8bits_to_6bits(pred, temp0, temp1, temp2, temp3, &output0, &output1, &output2);
st_dec_output = svcreate3_u8(output0, output1, output2);
svst3_u8(pred, (uint8_t *)*dst, st_dec_output);
pred_count = svcntp_b8(pred, pred);
*src += pred_count * 4;
*dst += pred_count * 3;
*src_len -= pred_count * 4;
i += pred_count * 4;
}
} else if (vl >= 128) {
pred = svwhilelt_b8((uint64_t)0, (uint64_t)128);
tbl_dec0 = svld1_u8(pred, (uint8_t *)dec_table + 0);
while (i < N) {
pred = svwhilelt_b8(i / 4, N / 4);
ld_dec_input = svld4_u8(pred, *src);
str0 = svget4_u8(ld_dec_input, 0);
str1 = svget4_u8(ld_dec_input, 1);
str2 = svget4_u8(ld_dec_input, 2);
str3 = svget4_u8(ld_dec_input, 3);
if (svcntp_b8(pred, svcmpge_n_u8(pred, str0, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str1, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str2, 128))) { return; }
if (svcntp_b8(pred, svcmpge_n_u8(pred, str3, 128))) { return; }
temp0 = svtbl_u8(tbl_dec0, str0);
temp1 = svtbl_u8(tbl_dec0, str1);
temp2 = svtbl_u8(tbl_dec0, str2);
temp3 = svtbl_u8(tbl_dec0, str3);
if (svcntp_b8(pred, svcmpeq_n_u8(pred, temp0, 255))) { return; }
if (svcntp_b8(pred, svcmpeq_n_u8(pred, temp1, 255))) { return; }
if (svcntp_b8(pred, svcmpeq_n_u8(pred, temp2, 255))) { return; }
if (svcntp_b8(pred, svcmpeq_n_u8(pred, temp3, 255))) { return; }
convert_8bits_to_6bits(pred, temp0, temp1, temp2, temp3, &output0, &output1, &output2);
st_dec_output = svcreate3_u8(output0, output1, output2);
svst3_u8(pred, (uint8_t *)*dst, st_dec_output);
pred_count = svcntp_b8(pred, pred);
*src += pred_count * 4;
*dst += pred_count * 3;
*src_len -= pred_count * 4;
i += pred_count * 4;
}
}
}