blob: 6a6d75f4c3d9806d42f658716ab11ae700e06dc3 [file] [log] [blame]
// Copyright 2025 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include <riscv_vector.h>
#include <stdint.h>
namespace {
constexpr size_t lut_size = 256;
// Double sized so we can check trailing regions are not read/written.
constexpr size_t buf_size = 256;
} // namespace
size_t vl __attribute__((section(".data"))) = 16;
// Indices are always unsigned.
uint8_t index_buf[buf_size] __attribute__((section(".data")));
// These instructions don't differentiate signed/unsigned so we only need to
// test one. The types come from intrinsic level.
uint8_t in_buf[lut_size] __attribute__((section(".data")));
uint8_t out_buf[buf_size] __attribute__((section(".data")));
extern "C" {
// Unordered
__attribute__((used, retain)) void vluxei8_v_u8mf4() {
auto indices = __riscv_vle8_v_u8mf4(index_buf, vl);
auto data = __riscv_vluxei8_v_u8mf4(in_buf, indices, vl);
__riscv_vse8_v_u8mf4(out_buf, data, vl);
}
__attribute__((used, retain)) void vluxei8_v_u8mf2() {
auto indices = __riscv_vle8_v_u8mf2(index_buf, vl);
auto data = __riscv_vluxei8_v_u8mf2(in_buf, indices, vl);
__riscv_vse8_v_u8mf2(out_buf, data, vl);
}
__attribute__((used, retain)) void vluxei8_v_u8m1() {
auto indices = __riscv_vle8_v_u8m1(index_buf, vl);
auto data = __riscv_vluxei8_v_u8m1(in_buf, indices, vl);
__riscv_vse8_v_u8m1(out_buf, data, vl);
}
__attribute__((used, retain)) void vluxei8_v_u8m2() {
auto indices = __riscv_vle8_v_u8m2(index_buf, vl);
auto data = __riscv_vluxei8_v_u8m2(in_buf, indices, vl);
__riscv_vse8_v_u8m2(out_buf, data, vl);
}
__attribute__((used, retain)) void vluxei8_v_u8m4() {
auto indices = __riscv_vle8_v_u8m4(index_buf, vl);
auto data = __riscv_vluxei8_v_u8m4(in_buf, indices, vl);
__riscv_vse8_v_u8m4(out_buf, data, vl);
}
__attribute__((used, retain)) void vluxei8_v_u8m8() {
auto indices = __riscv_vle8_v_u8m8(index_buf, vl);
auto data = __riscv_vluxei8_v_u8m8(in_buf, indices, vl);
__riscv_vse8_v_u8m8(out_buf, data, vl);
}
// Ordered
__attribute__((used, retain)) void vloxei8_v_u8mf4() {
auto indices = __riscv_vle8_v_u8mf4(index_buf, vl);
auto data = __riscv_vloxei8_v_u8mf4(in_buf, indices, vl);
__riscv_vse8_v_u8mf4(out_buf, data, vl);
}
__attribute__((used, retain)) void vloxei8_v_u8mf2() {
auto indices = __riscv_vle8_v_u8mf2(index_buf, vl);
auto data = __riscv_vloxei8_v_u8mf2(in_buf, indices, vl);
__riscv_vse8_v_u8mf2(out_buf, data, vl);
}
__attribute__((used, retain)) void vloxei8_v_u8m1() {
auto indices = __riscv_vle8_v_u8m1(index_buf, vl);
auto data = __riscv_vloxei8_v_u8m1(in_buf, indices, vl);
__riscv_vse8_v_u8m1(out_buf, data, vl);
}
__attribute__((used, retain)) void vloxei8_v_u8m2() {
auto indices = __riscv_vle8_v_u8m2(index_buf, vl);
auto data = __riscv_vloxei8_v_u8m2(in_buf, indices, vl);
__riscv_vse8_v_u8m2(out_buf, data, vl);
}
__attribute__((used, retain)) void vloxei8_v_u8m4() {
auto indices = __riscv_vle8_v_u8m4(index_buf, vl);
auto data = __riscv_vloxei8_v_u8m4(in_buf, indices, vl);
__riscv_vse8_v_u8m4(out_buf, data, vl);
}
__attribute__((used, retain)) void vloxei8_v_u8m8() {
auto indices = __riscv_vle8_v_u8m8(index_buf, vl);
auto data = __riscv_vloxei8_v_u8m8(in_buf, indices, vl);
__riscv_vse8_v_u8m8(out_buf, data, vl);
}
}
void (*impl)() __attribute__((section(".data"))) = &vluxei8_v_u8m1;
int main(int argc, char** argv) {
impl();
return 0;
}