diff --git a/apps/riscv-tests/isa/rv64uv/vlseg.c b/apps/riscv-tests/isa/rv64uv/vlseg.c index 3686ddae5..5febd509c 100644 --- a/apps/riscv-tests/isa/rv64uv/vlseg.c +++ b/apps/riscv-tests/isa/rv64uv/vlseg.c @@ -14,8 +14,8 @@ void TEST_CASE1_8(void) { VCLEAR(v1); VCLEAR(v2); asm volatile("vlseg2e8.v v1, (%0)" ::"r"(INP1)); - VCMP_U8(3, v1, 0x9f, 0x19, 0x8f, 0x05); - VCMP_U8(3, v2, 0xe4, 0x20, 0x2e, 0xe0); + VCMP_U8(1, v1, 0x9f, 0x19, 0x8f, 0x05); + VCMP_U8(2, v2, 0xe4, 0x20, 0x2e, 0xe0); } // Segment-3 @@ -27,9 +27,9 @@ void TEST_CASE2_8(void) { VCLEAR(v2); VCLEAR(v3); asm volatile("vlseg3e8.v v1, (%0)" ::"r"(INP1)); - VCMP_U8(4, v1, 0x9f, 0x20, 0x05, 0xaa); + VCMP_U8(3, v1, 0x9f, 0x20, 0x05, 0xaa); VCMP_U8(4, v2, 0xe4, 0x8f, 0xe0, 0x71); - VCMP_U8(4, v3, 0x19, 0x2e, 0xf9, 0xf0); + VCMP_U8(5, v3, 0x19, 0x2e, 0xf9, 0xf0); } // Segment-4 @@ -42,10 +42,10 @@ void TEST_CASE3_8(void) { VCLEAR(v3); VCLEAR(v4); asm volatile("vlseg4e8.v v1, (%0)" ::"r"(INP1)); - VCMP_U8(5, v1, 0x9f, 0x8f, 0xf9, 0xc3); - VCMP_U8(5, v2, 0xe4, 0x2e, 0xaa, 0x94); - VCMP_U8(5, v3, 0x19, 0x05, 0x71, 0xbb); - VCMP_U8(5, v4, 0x20, 0xe0, 0xf0, 0xd3); + VCMP_U8(6, v1, 0x9f, 0x8f, 0xf9, 0xc3); + VCMP_U8(7, v2, 0xe4, 0x2e, 0xaa, 0x94); + VCMP_U8(8, v3, 0x19, 0x05, 0x71, 0xbb); + VCMP_U8(9, v4, 0x20, 0xe0, 0xf0, 0xd3); } // Segment-8 @@ -62,14 +62,14 @@ void TEST_CASE4_8(void) { VCLEAR(v7); VCLEAR(v8); asm volatile("vlseg8e8.v v1, (%0)" ::"r"(INP1)); - VCMP_U8(6, v1, 0x9f, 0xf9); - VCMP_U8(6, v2, 0xe4, 0xaa); - VCMP_U8(6, v3, 0x19, 0x71); - VCMP_U8(6, v4, 0x20, 0xf0); - VCMP_U8(6, v5, 0x8f, 0xc3); - VCMP_U8(6, v6, 0x2e, 0x94); - VCMP_U8(6, v7, 0x05, 0xbb); - VCMP_U8(6, v8, 0xe0, 0xd3); + VCMP_U8(10, v1, 0x9f, 0xf9); + VCMP_U8(11, v2, 0xe4, 0xaa); + VCMP_U8(12, v3, 0x19, 0x71); + VCMP_U8(13, v4, 0x20, 0xf0); + VCMP_U8(14, v5, 0x8f, 0xc3); + VCMP_U8(15, v6, 0x2e, 0x94); + VCMP_U8(16, v7, 0x05, 0xbb); + VCMP_U8(17, v8, 0xe0, 0xd3); } // Segment-2 for 16-bit @@ -82,8 +82,8 @@ void TEST_CASE1_16(void) { VCLEAR(v1); VCLEAR(v2); asm volatile("vlseg2e16.v v1, (%0)" ::"r"(INP1)); - VCMP_U16(3, v1, 0x9fe4, 0x8f2e, 0xf9aa, 0xc394); - VCMP_U16(3, v2, 0x1920, 0x05e0, 0x71f0, 0xbbd3); + VCMP_U16(18, v1, 0x9fe4, 0x8f2e, 0xf9aa, 0xc394); + VCMP_U16(19, v2, 0x1920, 0x05e0, 0x71f0, 0xbbd3); } // Segment-3 for 16-bit @@ -97,9 +97,9 @@ void TEST_CASE2_16(void) { VCLEAR(v2); VCLEAR(v3); asm volatile("vlseg3e16.v v1, (%0)" ::"r"(INP1)); - VCMP_U16(4, v1, 0x9fe4, 0x05e0, 0xc394, 0x5678); - VCMP_U16(4, v2, 0x1920, 0xf9aa, 0xbbd3, 0x9abc); - VCMP_U16(4, v3, 0x8f2e, 0x71f0, 0x1234, 0xdef0); + VCMP_U16(20, v1, 0x9fe4, 0x05e0, 0xc394, 0x5678); + VCMP_U16(21, v2, 0x1920, 0xf9aa, 0xbbd3, 0x9abc); + VCMP_U16(22, v3, 0x8f2e, 0x71f0, 0x1234, 0xdef0); } // Segment-4 for 16-bit @@ -114,10 +114,10 @@ void TEST_CASE3_16(void) { VCLEAR(v3); VCLEAR(v4); asm volatile("vlseg4e16.v v1, (%0)" ::"r"(INP1)); - VCMP_U16(5, v1, 0x9fe4, 0xf9aa, 0x1234, 0x1357); - VCMP_U16(5, v2, 0x1920, 0x71f0, 0x5678, 0x2468); - VCMP_U16(5, v3, 0x8f2e, 0xc394, 0x9abc, 0x369b); - VCMP_U16(5, v4, 0x05e0, 0xbbd3, 0xdef0, 0x48ac); + VCMP_U16(23, v1, 0x9fe4, 0xf9aa, 0x1234, 0x1357); + VCMP_U16(24, v2, 0x1920, 0x71f0, 0x5678, 0x2468); + VCMP_U16(25, v3, 0x8f2e, 0xc394, 0x9abc, 0x369b); + VCMP_U16(26, v4, 0x05e0, 0xbbd3, 0xdef0, 0x48ac); } // Segment-8 for 16-bit @@ -136,14 +136,14 @@ void TEST_CASE4_16(void) { VCLEAR(v7); VCLEAR(v8); asm volatile("vlseg8e16.v v1, (%0)" ::"r"(INP1)); - VCMP_U16(6, v1, 0x9fe4, 0x1234); - VCMP_U16(6, v2, 0x1920, 0x5678); - VCMP_U16(6, v3, 0x8f2e, 0x9abc); - VCMP_U16(6, v4, 0x05e0, 0xdef0); - VCMP_U16(6, v5, 0xf9aa, 0x1357); - VCMP_U16(6, v6, 0x71f0, 0x2468); - VCMP_U16(6, v7, 0xc394, 0x369b); - VCMP_U16(6, v8, 0xbbd3, 0x48ac); + VCMP_U16(27, v1, 0x9fe4, 0x1234); + VCMP_U16(28, v2, 0x1920, 0x5678); + VCMP_U16(29, v3, 0x8f2e, 0x9abc); + VCMP_U16(30, v4, 0x05e0, 0xdef0); + VCMP_U16(31, v5, 0xf9aa, 0x1357); + VCMP_U16(32, v6, 0x71f0, 0x2468); + VCMP_U16(33, v7, 0xc394, 0x369b); + VCMP_U16(34, v8, 0xbbd3, 0x48ac); } // Segment-2 for 32-bit @@ -156,8 +156,8 @@ void TEST_CASE1_32(void) { VCLEAR(v1); VCLEAR(v2); asm volatile("vlseg2e32.v v1, (%0)" ::"r"(INP1)); - VCMP_U32(3, v1, 0x9fe41920, 0xf9aa71f0, 0x12345678, 0x13572468); - VCMP_U32(3, v2, 0x8f2e05e0, 0xc394bbd3, 0x9abcdef0, 0x369b48ac); + VCMP_U32(35, v1, 0x9fe41920, 0xf9aa71f0, 0x12345678, 0x13572468); + VCMP_U32(36, v2, 0x8f2e05e0, 0xc394bbd3, 0x9abcdef0, 0x369b48ac); } // Segment-3 for 32-bit @@ -171,9 +171,9 @@ void TEST_CASE2_32(void) { VCLEAR(v2); VCLEAR(v3); asm volatile("vlseg3e32.v v1, (%0)" ::"r"(INP1)); - VCMP_U32(4, v1, 0x9fe41920, 0xc394bbd3, 0x13572468, 0xcafebabe); - VCMP_U32(4, v2, 0x8f2e05e0, 0x12345678, 0x369b48ac, 0x01234567); - VCMP_U32(4, v3, 0xf9aa71f0, 0x9abcdef0, 0xdeadbeef, 0x89abcdef); + VCMP_U32(37, v1, 0x9fe41920, 0xc394bbd3, 0x13572468, 0xcafebabe); + VCMP_U32(38, v2, 0x8f2e05e0, 0x12345678, 0x369b48ac, 0x01234567); + VCMP_U32(39, v3, 0xf9aa71f0, 0x9abcdef0, 0xdeadbeef, 0x89abcdef); } // Segment-4 for 32-bit @@ -188,10 +188,10 @@ void TEST_CASE3_32(void) { VCLEAR(v3); VCLEAR(v4); asm volatile("vlseg4e32.v v1, (%0)" ::"r"(INP1)); - VCMP_U32(5, v1, 0x9fe41920, 0x12345678, 0xdeadbeef, 0x55aa55aa); - VCMP_U32(5, v2, 0x8f2e05e0, 0x9abcdef0, 0xcafebabe, 0x77889900); - VCMP_U32(5, v3, 0xf9aa71f0, 0x13572468, 0x01234567, 0xabcdef12); - VCMP_U32(5, v4, 0xc394bbd3, 0x369b48ac, 0x89abcdef, 0x34567890); + VCMP_U32(40, v1, 0x9fe41920, 0x12345678, 0xdeadbeef, 0x55aa55aa); + VCMP_U32(41, v2, 0x8f2e05e0, 0x9abcdef0, 0xcafebabe, 0x77889900); + VCMP_U32(42, v3, 0xf9aa71f0, 0x13572468, 0x01234567, 0xabcdef12); + VCMP_U32(43, v4, 0xc394bbd3, 0x369b48ac, 0x89abcdef, 0x34567890); } // Segment-8 for 32-bit @@ -210,14 +210,14 @@ void TEST_CASE4_32(void) { VCLEAR(v7); VCLEAR(v8); asm volatile("vlseg8e32.v v1, (%0)" ::"r"(INP1)); - VCMP_U32(6, v1, 0x9fe41920, 0xdeadbeef); - VCMP_U32(6, v2, 0x8f2e05e0, 0xcafebabe); - VCMP_U32(6, v3, 0xf9aa71f0, 0x01234567); - VCMP_U32(6, v4, 0xc394bbd3, 0x89abcdef); - VCMP_U32(6, v5, 0x12345678, 0x55aa55aa); - VCMP_U32(6, v6, 0x9abcdef0, 0x77889900); - VCMP_U32(6, v7, 0x13572468, 0xabcdef12); - VCMP_U32(6, v8, 0x369b48ac, 0x34567890); + VCMP_U32(44, v1, 0x9fe41920, 0xdeadbeef); + VCMP_U32(45, v2, 0x8f2e05e0, 0xcafebabe); + VCMP_U32(46, v3, 0xf9aa71f0, 0x01234567); + VCMP_U32(47, v4, 0xc394bbd3, 0x89abcdef); + VCMP_U32(48, v5, 0x12345678, 0x55aa55aa); + VCMP_U32(49, v6, 0x9abcdef0, 0x77889900); + VCMP_U32(50, v7, 0x13572468, 0xabcdef12); + VCMP_U32(51, v8, 0x369b48ac, 0x34567890); } // Segment-2 for 64-bit @@ -234,8 +234,8 @@ void TEST_CASE1_64(void) { VCLEAR(v1); VCLEAR(v2); asm volatile("vlseg2e64.v v1, (%0)" ::"r"(INP1)); - VCMP_U64(3, v1, 0x9fe419208f2e05e0, 0x123456789abcdef0, 0xdeadbeefcafebabe, 0x55aa55aa77889900); - VCMP_U64(3, v2, 0xf9aa71f0c394bbd3, 0x13572468369b48ac, 0x0123456789abcdef, 0xabcdef1234567890); + VCMP_U64(52, v1, 0x9fe419208f2e05e0, 0x123456789abcdef0, 0xdeadbeefcafebabe, 0x55aa55aa77889900); + VCMP_U64(53, v2, 0xf9aa71f0c394bbd3, 0x13572468369b48ac, 0x0123456789abcdef, 0xabcdef1234567890); } // Segment-3 for 64-bit @@ -253,9 +253,9 @@ void TEST_CASE2_64(void) { VCLEAR(v2); VCLEAR(v3); asm volatile("vlseg3e64.v v1, (%0)" ::"r"(INP1)); - VCMP_U64(4, v1, 0x9fe419208f2e05e0, 0x13572468369b48ac, 0x55aa55aa77889900, 0x123456789abcdef0); - VCMP_U64(4, v2, 0xf9aa71f0c394bbd3, 0xdeadbeefcafebabe, 0xabcdef1234567890, 0x1357246855aa55aa); - VCMP_U64(4, v3, 0x123456789abcdef0, 0x0123456789abcdef, 0xfeedfacecafebabe, 0x369b48acdeadbeef); + VCMP_U64(54, v1, 0x9fe419208f2e05e0, 0x13572468369b48ac, 0x55aa55aa77889900, 0x123456789abcdef0); + VCMP_U64(55, v2, 0xf9aa71f0c394bbd3, 0xdeadbeefcafebabe, 0xabcdef1234567890, 0x1357246855aa55aa); + VCMP_U64(56, v3, 0x123456789abcdef0, 0x0123456789abcdef, 0xfeedfacecafebabe, 0x369b48acdeadbeef); } // Segment-4 for 64-bit @@ -274,10 +274,10 @@ void TEST_CASE3_64(void) { VCLEAR(v3); VCLEAR(v4); asm volatile("vlseg4e64.v v1, (%0)" ::"r"(INP1)); - VCMP_U64(5, v1, 0x9fe419208f2e05e0, 0xdeadbeefcafebabe, 0xfeedfacecafebabe, 0xcafebabe12345678); - VCMP_U64(5, v2, 0xf9aa71f0c394bbd3, 0x0123456789abcdef, 0x123456789abcdef0, 0xabcdef0987654321); - VCMP_U64(5, v3, 0x123456789abcdef0, 0x55aa55aa77889900, 0x1357246855aa55aa, 0x012345670abcdef1); - VCMP_U64(5, v4, 0x13572468369b48ac, 0xabcdef1234567890, 0x369b48acdeadbeef, 0x987654321fedcba0); + VCMP_U64(57, v1, 0x9fe419208f2e05e0, 0xdeadbeefcafebabe, 0xfeedfacecafebabe, 0xcafebabe12345678); + VCMP_U64(58, v2, 0xf9aa71f0c394bbd3, 0x0123456789abcdef, 0x123456789abcdef0, 0xabcdef0987654321); + VCMP_U64(59, v3, 0x123456789abcdef0, 0x55aa55aa77889900, 0x1357246855aa55aa, 0x012345670abcdef1); + VCMP_U64(60, v4, 0x13572468369b48ac, 0xabcdef1234567890, 0x369b48acdeadbeef, 0x987654321fedcba0); } // Segment-8 for 64-bit @@ -300,14 +300,14 @@ void TEST_CASE4_64(void) { VCLEAR(v7); VCLEAR(v8); asm volatile("vlseg8e64.v v1, (%0)" ::"r"(INP1)); - VCMP_U64(6, v1, 0x9fe419208f2e05e0, 0xfeedfacecafebabe); - VCMP_U64(6, v2, 0xf9aa71f0c394bbd3, 0x123456789abcdef0); - VCMP_U64(6, v3, 0x123456789abcdef0, 0x1357246855aa55aa); - VCMP_U64(6, v4, 0x13572468369b48ac, 0x369b48acdeadbeef); - VCMP_U64(6, v5, 0xdeadbeefcafebabe, 0xcafebabe12345678); - VCMP_U64(6, v6, 0x0123456789abcdef, 0xabcdef0987654321); - VCMP_U64(6, v7, 0x55aa55aa77889900, 0x012345670abcdef1); - VCMP_U64(6, v8, 0xabcdef1234567890, 0x987654321fedcba0); + VCMP_U64(61, v1, 0x9fe419208f2e05e0, 0xfeedfacecafebabe); + VCMP_U64(62, v2, 0xf9aa71f0c394bbd3, 0x123456789abcdef0); + VCMP_U64(63, v3, 0x123456789abcdef0, 0x1357246855aa55aa); + VCMP_U64(64, v4, 0x13572468369b48ac, 0x369b48acdeadbeef); + VCMP_U64(65, v5, 0xdeadbeefcafebabe, 0xcafebabe12345678); + VCMP_U64(66, v6, 0x0123456789abcdef, 0xabcdef0987654321); + VCMP_U64(67, v7, 0x55aa55aa77889900, 0x012345670abcdef1); + VCMP_U64(68, v8, 0xabcdef1234567890, 0x987654321fedcba0); } // Segment-8 for 64-bit @@ -331,14 +331,14 @@ void TEST_CASE4_64_m(void) { VCLEAR(v8); VLOAD_8(v0, 0xAA, 0xAA); asm volatile("vlseg8e64.v v1, (%0), v0.t" ::"r"(INP1)); - VCMP_U64(6, v1, 0, 0xfeedfacecafebabe); - VCMP_U64(6, v2, 0, 0x123456789abcdef0); - VCMP_U64(6, v3, 0, 0x1357246855aa55aa); - VCMP_U64(6, v4, 0, 0x369b48acdeadbeef); - VCMP_U64(6, v5, 0, 0xcafebabe12345678); - VCMP_U64(6, v6, 0, 0xabcdef0987654321); - VCMP_U64(6, v7, 0, 0x012345670abcdef1); - VCMP_U64(6, v8, 0, 0x987654321fedcba0); + VCMP_U64(69, v1, 0, 0xfeedfacecafebabe); + VCMP_U64(70, v2, 0, 0x123456789abcdef0); + VCMP_U64(71, v3, 0, 0x1357246855aa55aa); + VCMP_U64(72, v4, 0, 0x369b48acdeadbeef); + VCMP_U64(73, v5, 0, 0xcafebabe12345678); + VCMP_U64(74, v6, 0, 0xabcdef0987654321); + VCMP_U64(75, v7, 0, 0x012345670abcdef1); + VCMP_U64(76, v8, 0, 0x987654321fedcba0); VCLEAR(v1); VCLEAR(v2); @@ -350,14 +350,14 @@ void TEST_CASE4_64_m(void) { VCLEAR(v8); VLOAD_8(v0, 0xA8, 0xAA); asm volatile("vlseg8e64.v v1, (%0), v0.t" ::"r"(INP1)); - VCMP_U64(6, v1, 0, 0); - VCMP_U64(6, v2, 0, 0); - VCMP_U64(6, v3, 0, 0); - VCMP_U64(6, v4, 0, 0); - VCMP_U64(6, v5, 0, 0); - VCMP_U64(6, v6, 0, 0); - VCMP_U64(6, v7, 0, 0); - VCMP_U64(6, v8, 0, 0); + VCMP_U64(77, v1, 0, 0); + VCMP_U64(78, v2, 0, 0); + VCMP_U64(79, v3, 0, 0); + VCMP_U64(80, v4, 0, 0); + VCMP_U64(81, v5, 0, 0); + VCMP_U64(82, v6, 0, 0); + VCMP_U64(83, v7, 0, 0); + VCMP_U64(84, v8, 0, 0); }