blob: 7338eb21b67d651c13c579ebe07acaac50d1fa19 [file] [log] [blame]
// Copyright (c) 2017-2019 ARM Limited
// All rights reserved
//
// The license below extends only to copyright in the software and shall
// not be construed as granting a license to any other intellectual
// property including but not limited to intellectual property relating
// to a hardware implementation of the functionality of the software
// licensed hereunder. You may use the software subject to the license
// terms below provided that you ensure that this notice is replicated
// unmodified and in its entirety in all distributions of the software,
// modified or unmodified, in source code or in binary form.
//
// Redistribution and use in source and binary forms, with or without
// modification, are permitted provided that the following conditions are
// met: redistributions of source code must retain the above copyright
// notice, this list of conditions and the following disclaimer;
// redistributions in binary form must reproduce the above copyright
// notice, this list of conditions and the following disclaimer in the
// documentation and/or other materials provided with the distribution;
// neither the name of the copyright holders nor the names of its
// contributors may be used to endorse or promote products derived from
// this software without specific prior written permission.
//
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
// @file Definition of SVE instructions.
output header {{
// Decodes unary, constructive, predicated (merging) SVE instructions,
// handling signed and unsigned variants.
template <template <typename T> class BaseS,
template <typename T> class BaseU>
StaticInstPtr
decodeSveUnaryPred(unsigned size, unsigned u, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex gp)
{
switch (size) {
case 0:
if (u) {
return new BaseU<uint8_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int8_t>(machInst, dest, op1, gp);
}
case 1:
if (u) {
return new BaseU<uint16_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int16_t>(machInst, dest, op1, gp);
}
case 2:
if (u) {
return new BaseU<uint32_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int32_t>(machInst, dest, op1, gp);
}
case 3:
if (u) {
return new BaseU<uint64_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int64_t>(machInst, dest, op1, gp);
}
default:
return new Unknown64(machInst);
}
}
// Decodes SVE widening reductions.
// handling signed and unsigned variants.
template <template <typename T1, typename T2> class BaseS,
template <typename T1, typename T2> class BaseU>
StaticInstPtr
decodeSveWideningReduc(unsigned size, unsigned u, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex gp)
{
switch (size) {
case 0:
if (u) {
return new BaseU<uint8_t, uint64_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int8_t, int64_t>(machInst, dest, op1, gp);
}
case 1:
if (u) {
return new BaseU<uint16_t, uint64_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int16_t, int64_t>(machInst, dest, op1, gp);
}
case 2:
if (u) {
return new BaseU<uint32_t, uint64_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int32_t, int64_t>(machInst, dest, op1, gp);
}
case 3:
assert(u);
return new BaseU<uint64_t, uint64_t>(machInst, dest, op1, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes unary, constructive, predicated (merging) SVE instructions,
// handling signed variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveUnaryPredS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, op1, gp);
case 1:
return new Base<int16_t>(machInst, dest, op1, gp);
case 2:
return new Base<int32_t>(machInst, dest, op1, gp);
case 3:
return new Base<int64_t>(machInst, dest, op1, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes unary, constructive, predicated (merging) SVE instructions,
// handling unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveUnaryPredU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op1, gp);
case 1:
return new Base<uint16_t>(machInst, dest, op1, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op1, gp);
case 3:
return new Base<uint64_t>(machInst, dest, op1, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes unary, constructive, predicated (merging) SVE instructions,
// handling signed and unsigned variants, for small element sizes (8- to
// 32-bit).
template <template <typename T> class BaseS,
template <typename T> class BaseU>
StaticInstPtr
decodeSveUnaryPredSmall(unsigned size, unsigned u, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex gp)
{
switch (size) {
case 0:
if (u) {
return new BaseU<uint8_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int8_t>(machInst, dest, op1, gp);
}
case 1:
if (u) {
return new BaseU<uint16_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int16_t>(machInst, dest, op1, gp);
}
case 2:
if (u) {
return new BaseU<uint32_t>(machInst, dest, op1, gp);
} else {
return new BaseS<int32_t>(machInst, dest, op1, gp);
}
default:
return new Unknown64(machInst);
}
}
// Decodes unary, constructive, predicated (merging) SVE instructions,
// handling floating point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveUnaryPredF(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex gp)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, op1, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op1, gp);
case 3:
return new Base<uint64_t>(machInst, dest, op1, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes unary, constructive, unpredicated SVE instructions, handling
// unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveUnaryUnpredU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op1);
case 1:
return new Base<uint16_t>(machInst, dest, op1);
case 2:
return new Base<uint32_t>(machInst, dest, op1);
case 3:
return new Base<uint64_t>(machInst, dest, op1);
default:
return new Unknown64(machInst);
}
}
// Decodes unary, constructive, unpredicated SVE instructions, handling
// floating-point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveUnaryUnpredF(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, op1);
case 2:
return new Base<uint32_t>(machInst, dest, op1);
case 3:
return new Base<uint64_t>(machInst, dest, op1);
default:
return new Unknown64(machInst);
}
}
// Decodes binary, destructive, predicated (merging) SVE instructions,
// handling signed and unsigned variants.
template <template <typename T> class BaseS,
template <typename T> class BaseU>
StaticInstPtr
decodeSveBinDestrPred(unsigned size, unsigned u, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op2, IntRegIndex gp)
{
switch (size) {
case 0:
if (u) {
return new BaseU<uint8_t>(machInst, dest, op2, gp);
} else {
return new BaseS<int8_t>(machInst, dest, op2, gp);
}
case 1:
if (u) {
return new BaseU<uint16_t>(machInst, dest, op2, gp);
} else {
return new BaseS<int16_t>(machInst, dest, op2, gp);
}
case 2:
if (u) {
return new BaseU<uint32_t>(machInst, dest, op2, gp);
} else {
return new BaseS<int32_t>(machInst, dest, op2, gp);
}
case 3:
if (u) {
return new BaseU<uint64_t>(machInst, dest, op2, gp);
} else {
return new BaseS<int64_t>(machInst, dest, op2, gp);
}
default:
return new Unknown64(machInst);
}
}
// Decodes binary with immediate operand, constructive, unpredicated
// SVE instructions, handling signed variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinImmUnpredS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, unsigned immediate)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, op1, immediate);
case 1:
return new Base<int16_t>(machInst, dest, op1, immediate);
case 2:
return new Base<int32_t>(machInst, dest, op1, immediate);
case 3:
return new Base<int64_t>(machInst, dest, op1, immediate);
default:
return new Unknown64(machInst);
}
}
// Decodes binary with immediate operand, constructive, unpredicated
// SVE instructions, handling unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinImmUnpredU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, unsigned immediate)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op1, immediate);
case 1:
return new Base<uint16_t>(machInst, dest, op1, immediate);
case 2:
return new Base<uint32_t>(machInst, dest, op1, immediate);
case 3:
return new Base<uint64_t>(machInst, dest, op1, immediate);
default:
return new Unknown64(machInst);
}
}
// Decodes binary with immediate operand, destructive, predicated (merging)
// SVE instructions, handling unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinImmPredU(unsigned size, ExtMachInst machInst, IntRegIndex dest,
unsigned immediate, IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, immediate, gp);
case 1:
return new Base<uint16_t>(machInst, dest, immediate, gp);
case 2:
return new Base<uint32_t>(machInst, dest, immediate, gp);
case 3:
return new Base<uint64_t>(machInst, dest, immediate, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes binary with immediate operand, destructive, predicated (merging)
// SVE instructions, handling signed variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinImmPredS(unsigned size, ExtMachInst machInst, IntRegIndex dest,
unsigned immediate, IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, immediate, gp);
case 1:
return new Base<int16_t>(machInst, dest, immediate, gp);
case 2:
return new Base<int32_t>(machInst, dest, immediate, gp);
case 3:
return new Base<int64_t>(machInst, dest, immediate, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes binary with immediate operand, destructive, predicated (merging)
// SVE instructions, handling floating-point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinImmPredF(unsigned size, ExtMachInst machInst, IntRegIndex dest,
uint64_t immediate, IntRegIndex gp)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, immediate, gp);
case 2:
return new Base<uint32_t>(machInst, dest, immediate, gp);
case 3:
return new Base<uint64_t>(machInst, dest, immediate, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes unary/binary with wide immediate operand, destructive,
// unpredicated SVE instructions, handling unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveWideImmUnpredU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint64_t immediate)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, immediate);
case 1:
return new Base<uint16_t>(machInst, dest, immediate);
case 2:
return new Base<uint32_t>(machInst, dest, immediate);
case 3:
return new Base<uint64_t>(machInst, dest, immediate);
default:
return new Unknown64(machInst);
}
}
// Decodes unary/binary with wide immediate operand, destructive,
// unpredicated SVE instructions, handling signed variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveWideImmUnpredS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint64_t immediate)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, immediate);
case 1:
return new Base<int16_t>(machInst, dest, immediate);
case 2:
return new Base<int32_t>(machInst, dest, immediate);
case 3:
return new Base<int64_t>(machInst, dest, immediate);
default:
return new Unknown64(machInst);
}
}
// Decodes unary/binary with wide immediate operand, destructive,
// unpredicated SVE instructions, handling floating-point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveWideImmUnpredF(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint64_t immediate)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, immediate);
case 2:
return new Base<uint32_t>(machInst, dest, immediate);
case 3:
return new Base<uint64_t>(machInst, dest, immediate);
default:
return new Unknown64(machInst);
}
}
// Decodes unary/binary with wide immediate operand, destructive,
// predicated SVE instructions, handling unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveWideImmPredU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint64_t immediate, IntRegIndex gp,
bool isMerging = true)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, immediate, gp,
isMerging);
case 1:
return new Base<uint16_t>(machInst, dest, immediate, gp,
isMerging);
case 2:
return new Base<uint32_t>(machInst, dest, immediate, gp,
isMerging);
case 3:
return new Base<uint64_t>(machInst, dest, immediate, gp,
isMerging);
default:
return new Unknown64(machInst);
}
}
// Decodes unary/binary with wide immediate operand, destructive,
// predicated SVE instructions, handling floating-point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveWideImmPredF(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint64_t immediate, IntRegIndex gp)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, immediate, gp);
case 2:
return new Base<uint32_t>(machInst, dest, immediate, gp);
case 3:
return new Base<uint64_t>(machInst, dest, immediate, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes binary, destructive, predicated (merging) SVE instructions,
// handling unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinDestrPredU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op2, IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op2, gp);
case 1:
return new Base<uint16_t>(machInst, dest, op2, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op2, gp);
case 3:
return new Base<uint64_t>(machInst, dest, op2, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes binary, destructive, predicated (merging) SVE instructions,
// handling signed variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinDestrPredS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op2, IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, op2, gp);
case 1:
return new Base<int16_t>(machInst, dest, op2, gp);
case 2:
return new Base<int32_t>(machInst, dest, op2, gp);
case 3:
return new Base<int64_t>(machInst, dest, op2, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes binary, destructive, predicated (merging) SVE instructions,
// handling floating-point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinDestrPredF(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op2, IntRegIndex gp)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, op2, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op2, gp);
case 3:
return new Base<uint64_t>(machInst, dest, op2, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes binary, constructive, predicated SVE instructions, handling
// unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinConstrPredU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex op2,
IntRegIndex gp, SvePredType predType)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op1, op2, gp, predType);
case 1:
return new Base<uint16_t>(machInst, dest, op1, op2, gp, predType);
case 2:
return new Base<uint32_t>(machInst, dest, op1, op2, gp, predType);
case 3:
return new Base<uint64_t>(machInst, dest, op1, op2, gp, predType);
default:
return new Unknown64(machInst);
}
}
// Decodes binary, constructive, unpredicated SVE instructions.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinUnpred(unsigned size, unsigned u, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex op2)
{
switch (size) {
case 0:
if (u) {
return new Base<uint8_t>(machInst, dest, op1, op2);
} else {
return new Base<int8_t>(machInst, dest, op1, op2);
}
case 1:
if (u) {
return new Base<uint16_t>(machInst, dest, op1, op2);
} else {
return new Base<int16_t>(machInst, dest, op1, op2);
}
case 2:
if (u) {
return new Base<uint32_t>(machInst, dest, op1, op2);
} else {
return new Base<int32_t>(machInst, dest, op1, op2);
}
case 3:
if (u) {
return new Base<uint64_t>(machInst, dest, op1, op2);
} else {
return new Base<int64_t>(machInst, dest, op1, op2);
}
default:
return new Unknown64(machInst);
}
}
// Decodes binary, constructive, unpredicated SVE instructions.
// Unsigned instructions only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinUnpredU(unsigned size, ExtMachInst machInst, IntRegIndex dest,
IntRegIndex op1, IntRegIndex op2)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op1, op2);
case 1:
return new Base<uint16_t>(machInst, dest, op1, op2);
case 2:
return new Base<uint32_t>(machInst, dest, op1, op2);
case 3:
return new Base<uint64_t>(machInst, dest, op1, op2);
default:
return new Unknown64(machInst);
}
}
// Decodes binary, constructive, unpredicated SVE instructions.
// Signed instructions only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinUnpredS(unsigned size, ExtMachInst machInst, IntRegIndex dest,
IntRegIndex op1, IntRegIndex op2)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, op1, op2);
case 1:
return new Base<int16_t>(machInst, dest, op1, op2);
case 2:
return new Base<int32_t>(machInst, dest, op1, op2);
case 3:
return new Base<int64_t>(machInst, dest, op1, op2);
default:
return new Unknown64(machInst);
}
}
// Decodes binary, costructive, unpredicated SVE instructions, handling
// floating-point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveBinUnpredF(unsigned size, ExtMachInst machInst, IntRegIndex dest,
IntRegIndex op1, IntRegIndex op2)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, op1, op2);
case 2:
return new Base<uint32_t>(machInst, dest, op1, op2);
case 3:
return new Base<uint64_t>(machInst, dest, op1, op2);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE compare instructions - binary, predicated (zeroing),
// generating a predicate - handling floating-point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveCmpF(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex op2,
IntRegIndex gp)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, op1, op2, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op1, op2, gp);
case 3:
return new Base<uint64_t>(machInst, dest, op1, op2, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE compare-with-immediate instructions - binary, predicated
// (zeroing), generating a predicate - handling floating-point variants
// only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveCmpImmF(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, uint64_t imm,
IntRegIndex gp)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, op1, imm, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op1, imm, gp);
case 3:
return new Base<uint64_t>(machInst, dest, op1, imm, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes ternary, destructive, predicated (merging) SVE instructions.
template <template <typename T> class Base>
StaticInstPtr
decodeSveTerPred(unsigned size, unsigned u, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex op2,
IntRegIndex gp)
{
switch (size) {
case 0:
if (u) {
return new Base<uint8_t>(machInst, dest, op1, op2, gp);
} else {
return new Base<int8_t>(machInst, dest, op1, op2, gp);
}
case 1:
if (u) {
return new Base<uint16_t>(machInst, dest, op1, op2, gp);
} else {
return new Base<int16_t>(machInst, dest, op1, op2, gp);
}
case 2:
if (u) {
return new Base<uint32_t>(machInst, dest, op1, op2, gp);
} else {
return new Base<int32_t>(machInst, dest, op1, op2, gp);
}
case 3:
if (u) {
return new Base<uint64_t>(machInst, dest, op1, op2, gp);
} else {
return new Base<int64_t>(machInst, dest, op1, op2, gp);
}
default:
return new Unknown64(machInst);
}
}
// Decodes ternary, destructive, predicated (merging) SVE instructions,
// handling wide signed variants only. XXX: zeroing for CMP instructions.
template <template <typename T> class Base>
StaticInstPtr
decodeSveTerPredWS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex op2,
IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, op1, op2, gp);
case 1:
return new Base<int16_t>(machInst, dest, op1, op2, gp);
case 2:
return new Base<int32_t>(machInst, dest, op1, op2, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes ternary, destructive, predicated (merging) SVE instructions,
// handling wide unsigned variants only. XXX: zeroing for CMP instructions.
template <template <typename T> class Base>
StaticInstPtr
decodeSveTerPredWU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex op2,
IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op1, op2, gp);
case 1:
return new Base<uint16_t>(machInst, dest, op1, op2, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op1, op2, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes ternary, destructive, predicated (merging) SVE instructions,
// handling signed variants only. XXX: zeroing for CMP instructions.
template <template <typename T> class Base>
StaticInstPtr
decodeSveTerPredS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex op2,
IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, op1, op2, gp);
case 1:
return new Base<int16_t>(machInst, dest, op1, op2, gp);
case 2:
return new Base<int32_t>(machInst, dest, op1, op2, gp);
case 3:
return new Base<int64_t>(machInst, dest, op1, op2, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes ternary, destructive, predicated (merging) SVE instructions,
// handling unsigned variants only. XXX: zeroing for CMP instructions.
template <template <typename T> class Base>
StaticInstPtr
decodeSveTerPredU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex op2,
IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op1, op2, gp);
case 1:
return new Base<uint16_t>(machInst, dest, op1, op2, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op1, op2, gp);
case 3:
return new Base<uint64_t>(machInst, dest, op1, op2, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE signed unary extension instructions (8-bit source element
// size)
template <template <typename TS, typename TD> class Base>
StaticInstPtr
decodeSveUnaryExtendFromBPredS(unsigned dsize, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1,
IntRegIndex gp)
{
switch (dsize) {
case 1:
return new Base<int8_t, int16_t>(machInst, dest, op1, gp);
case 2:
return new Base<int8_t, int32_t>(machInst, dest, op1, gp);
case 3:
return new Base<int8_t, int64_t>(machInst, dest, op1, gp);
}
return new Unknown64(machInst);
}
// Decodes SVE unsigned unary extension instructions (8-bit source element
// size)
template <template <typename TS, typename TD> class Base>
StaticInstPtr
decodeSveUnaryExtendFromBPredU(unsigned dsize, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1,
IntRegIndex gp)
{
switch (dsize) {
case 1:
return new Base<uint8_t, uint16_t>(machInst, dest, op1, gp);
case 2:
return new Base<uint8_t, uint32_t>(machInst, dest, op1, gp);
case 3:
return new Base<uint8_t, uint64_t>(machInst, dest, op1, gp);
}
return new Unknown64(machInst);
}
// Decodes SVE signed unary extension instructions (16-bit source element
// size)
template <template <typename TS, typename TD> class Base>
StaticInstPtr
decodeSveUnaryExtendFromHPredS(unsigned dsize, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1,
IntRegIndex gp)
{
switch (dsize) {
case 2:
return new Base<int16_t, int32_t>(machInst, dest, op1, gp);
case 3:
return new Base<int16_t, int64_t>(machInst, dest, op1, gp);
}
return new Unknown64(machInst);
}
// Decodes SVE unsigned unary extension instructions (16-bit source element
// size)
template <template <typename TS, typename TD> class Base>
StaticInstPtr
decodeSveUnaryExtendFromHPredU(unsigned dsize, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1,
IntRegIndex gp)
{
switch (dsize) {
case 2:
return new Base<uint16_t, uint32_t>(machInst, dest, op1, gp);
case 3:
return new Base<uint16_t, uint64_t>(machInst, dest, op1, gp);
}
return new Unknown64(machInst);
}
// Decodes ternary, destructive, predicated (merging) SVE instructions,
// handling floating-point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveTerPredF(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, IntRegIndex op2,
IntRegIndex gp)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, op1, op2, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op1, op2, gp);
case 3:
return new Base<uint64_t>(machInst, dest, op1, op2, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes ternary with immediate operand, destructive, unpredicated SVE
// instructions handling floating-point variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveTerImmUnpredF(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op2, uint8_t imm)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, op2, imm);
case 2:
return new Base<uint32_t>(machInst, dest, op2, imm);
case 3:
return new Base<uint64_t>(machInst, dest, op2, imm);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE PTRUE(S) instructions.
template <template <typename T> class Base>
StaticInstPtr
decodeSvePtrue(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint8_t imm)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, imm);
case 1:
return new Base<uint16_t>(machInst, dest, imm);
case 2:
return new Base<uint32_t>(machInst, dest, imm);
case 3:
return new Base<uint64_t>(machInst, dest, imm);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE predicate count instructions, scalar signed variant only
template <template <typename T> class Base>
StaticInstPtr
decodeSvePredCountS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, op1);
case 1:
return new Base<int16_t>(machInst, dest, op1);
case 2:
return new Base<int32_t>(machInst, dest, op1);
case 3:
return new Base<int64_t>(machInst, dest, op1);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE predicate count instructions, scalar unsigned variant only
template <template <typename T> class Base>
StaticInstPtr
decodeSvePredCountU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op1);
case 1:
return new Base<uint16_t>(machInst, dest, op1);
case 2:
return new Base<uint32_t>(machInst, dest, op1);
case 3:
return new Base<uint64_t>(machInst, dest, op1);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE predicate count instructions, vector signed variant only
template <template <typename T> class Base>
StaticInstPtr
decodeSvePredCountVS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1)
{
switch (size) {
case 1:
return new Base<int16_t>(machInst, dest, op1);
case 2:
return new Base<int32_t>(machInst, dest, op1);
case 3:
return new Base<int64_t>(machInst, dest, op1);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE predicate count instructions, vector unsigned variant only
template <template <typename T> class Base>
StaticInstPtr
decodeSvePredCountVU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, op1);
case 2:
return new Base<uint32_t>(machInst, dest, op1);
case 3:
return new Base<uint64_t>(machInst, dest, op1);
default:
return new Unknown64(machInst);
}
}
// Decodes ternary with immediate operand, predicated SVE
// instructions handling unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveTerImmPredU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, int64_t imm, IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, op1, imm, gp);
case 1:
return new Base<uint16_t>(machInst, dest, op1, imm, gp);
case 2:
return new Base<uint32_t>(machInst, dest, op1, imm, gp);
case 3:
return new Base<uint64_t>(machInst, dest, op1, imm, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes ternary with immediate operand, predicated SVE
// instructions handling signed variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveTerImmPredS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1, int64_t imm, IntRegIndex gp)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, op1, imm, gp);
case 1:
return new Base<int16_t>(machInst, dest, op1, imm, gp);
case 2:
return new Base<int32_t>(machInst, dest, op1, imm, gp);
case 3:
return new Base<int64_t>(machInst, dest, op1, imm, gp);
default:
return new Unknown64(machInst);
}
}
// Decodes integer element count SVE instructions, handling
// signed variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveElemIntCountS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint8_t pattern, uint8_t imm4)
{
switch (size) {
case 0:
return new Base<int8_t>(machInst, dest, pattern, imm4);
case 1:
return new Base<int16_t>(machInst, dest, pattern, imm4);
case 2:
return new Base<int32_t>(machInst, dest, pattern, imm4);
case 3:
return new Base<int64_t>(machInst, dest, pattern, imm4);
default:
return new Unknown64(machInst);
}
}
// Decodes integer element count SVE instructions, handling
// unsigned variants only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveElemIntCountU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint8_t pattern, uint8_t imm4)
{
switch (size) {
case 0:
return new Base<uint8_t>(machInst, dest, pattern, imm4);
case 1:
return new Base<uint16_t>(machInst, dest, pattern, imm4);
case 2:
return new Base<uint32_t>(machInst, dest, pattern, imm4);
case 3:
return new Base<uint64_t>(machInst, dest, pattern, imm4);
default:
return new Unknown64(machInst);
}
}
// Decodes integer element count SVE instructions, handling
// signed variants from 16 to 64 bits only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveElemIntCountLS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint8_t pattern, uint8_t imm4)
{
switch (size) {
case 1:
return new Base<int16_t>(machInst, dest, pattern, imm4);
case 2:
return new Base<int32_t>(machInst, dest, pattern, imm4);
case 3:
return new Base<int64_t>(machInst, dest, pattern, imm4);
default:
return new Unknown64(machInst);
}
}
// Decodes integer element count SVE instructions, handling
// unsigned variants from 16 to 64 bits only.
template <template <typename T> class Base>
StaticInstPtr
decodeSveElemIntCountLU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, uint8_t pattern, uint8_t imm4)
{
switch (size) {
case 1:
return new Base<uint16_t>(machInst, dest, pattern, imm4);
case 2:
return new Base<uint32_t>(machInst, dest, pattern, imm4);
case 3:
return new Base<uint64_t>(machInst, dest, pattern, imm4);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE unpack instructions. Handling signed variants.
template <template <typename T1, typename T2> class Base>
StaticInstPtr
decodeSveUnpackS(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1)
{
switch (size) {
case 1:
return new Base<int8_t, int16_t>(machInst, dest, op1);
case 2:
return new Base<int16_t, int32_t>(machInst, dest, op1);
case 3:
return new Base<int32_t, int64_t>(machInst, dest, op1);
default:
return new Unknown64(machInst);
}
}
// Decodes SVE unpack instructions. Handling unsigned variants.
template <template <typename T1, typename T2> class Base>
StaticInstPtr
decodeSveUnpackU(unsigned size, ExtMachInst machInst,
IntRegIndex dest, IntRegIndex op1)
{
switch (size) {
case 1:
return new Base<uint8_t, uint16_t>(machInst, dest, op1);
case 2:
return new Base<uint16_t, uint32_t>(machInst, dest, op1);
case 3:
return new Base<uint32_t, uint64_t>(machInst, dest, op1);
default:
return new Unknown64(machInst);
}
}
}};
let {{
header_output = ''
exec_output = ''
decoders = { 'Generic': {} }
class PredType:
NONE = 0
MERGE = 1
ZERO = 2
SELECT = 3
class CvtDir:
Narrow = 0
Widen = 1
class IndexFormat(object):
ImmImm = 'II'
ImmReg = 'IR'
RegImm = 'RI'
RegReg = 'RR'
class SrcRegType(object):
Vector = 0
Scalar = 1
SimdFpScalar = 2
Predicate = 3
class DstRegType(object):
Vector = 0
Scalar = 1
SimdFpScalar = 2
Predicate = 3
class DestType(object):
Scalar = 'false'
Vector = 'true'
class SrcSize(object):
Src32bit = 'true'
Src64bit = 'false'
class Break(object):
Before = 0
After = 1
class Unpack(object):
High = 0
Low = 1
# Generates definitions for SVE ADR instructions
def sveAdrInst(name, Name, opClass, types, op):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {
const Element& srcElem1 = AA64FpOp1_x[i];
Element srcElem2 = AA64FpOp2_x[i];
Element destElem = 0;
%(op)s
AA64FpDest_x[i] = destElem;
}''' % {'op': op}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveAdrOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveAdrOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definition for SVE while predicate generation instructions
def sveWhileInst(name, Name, opClass, types, op,
srcSize = SrcSize.Src64bit):
global header_output, exec_output, decoders
extraPrologCode = '''
auto& destPred = PDest;'''
if 'int32_t' in types:
srcType = 'int64_t' if srcSize == SrcSize.Src64bit else 'int32_t'
else:
srcType = 'uint64_t' if srcSize == SrcSize.Src64bit else 'uint32_t'
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
%(stype)s srcElem1 = static_cast<%(stype)s>(XOp1);
%(stype)s srcElem2 = static_cast<%(stype)s>(XOp2);
bool cond, first = false, none = true, last = true;
destPred.reset();
for (unsigned i = 0; i < eCount; i++) {
%(op)s;
last = last && cond;
none = none && !cond;
first = first || (i == 0 && cond);
PDest_x[i] = last;
srcElem1++;
}
CondCodesNZ = (first << 1) | none;
CondCodesC = !last;
CondCodesV = false;
'''%{'op': op, 'stype': srcType}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveWhileOp',
{'code': code, 'op_class': opClass, 'srcIs32b': srcSize}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SveWhileOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type, 'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict);
# Generate definition for SVE compare & terminate instructions
def sveCompTermInst(name, Name, opClass, types, op):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
bool destElem;
Element srcElem1 = static_cast<Element>(XOp1);
Element srcElem2 = static_cast<Element>(XOp2);
%(op)s;
if (destElem) {
CondCodesNZ = CondCodesNZ | 0x2;
CondCodesV = 0;
} else {
CondCodesNZ = CondCodesNZ & ~0x2;
CondCodesV = !CondCodesC;
}
''' % {'op': op}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveCompTermOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveCompTermOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type, 'class_name': 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict);
# Generates definition for SVE predicate count instructions
def svePredCountInst(name, Name, opClass, types, op,
destType=DestType.Vector,
srcSize=SrcSize.Src64bit):
global header_output, exec_output, decoders
assert not (destType == DestType.Vector and
srcSize != SrcSize.Src64bit)
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
int count = 0;
for (unsigned i = 0; i < eCount; i++) {
if (GpOp_x[i]) {
count++;
}
}'''
if destType == DestType.Vector:
code += '''
for (unsigned i = 0; i < eCount; i++) {
Element destElem = 0;
const Element& srcElem = AA64FpDestMerge_x[i];
%(op)s
AA64FpDest_x[i] = destElem;
}''' % {'op': op}
else:
code += '''
%(op)s''' % {'op': op}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SvePredCountOp',
{'code': code, 'op_class': opClass, 'srcIs32b': srcSize,
'destIsVec': destType}, [])
header_output += SvePredCountOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type, 'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict);
# Generates definition for SVE predicate count instructions (predicated)
def svePredCountPredInst(name, Name, opClass, types):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
int count = 0;
for (unsigned i = 0; i < eCount; i++) {
if (POp1_x[i] && GpOp_x[i]) {
count++;
}
}
XDest = count;
'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SvePredCountPredOp',
{'code': code, 'op_class': opClass}, [])
header_output += SvePredCountPredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type, 'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definition for SVE Index generation instructions
def sveIndex(fmt):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());'''
if fmt == IndexFormat.ImmReg or fmt == IndexFormat.ImmImm:
code += '''
const Element& srcElem1 = imm1;'''
if fmt == IndexFormat.RegImm or fmt == IndexFormat.RegReg:
code += '''
const Element& srcElem1 = XOp1;'''
if fmt == IndexFormat.RegImm or fmt == IndexFormat.ImmImm:
code += '''
const Element& srcElem2 = imm2;'''
if fmt == IndexFormat.ImmReg or fmt == IndexFormat.RegReg:
code += '''
const Element& srcElem2 = XOp2;'''
code +='''
for (unsigned i = 0; i < eCount; i++) {
AA64FpDest_x[i] = srcElem1 + i * srcElem2;
}'''
iop = ArmInstObjParams('index', 'SveIndex'+fmt, 'SveIndex'+fmt+'Op',
{'code': code, 'op_class': 'SimdAluOp'})
if fmt == IndexFormat.ImmImm:
header_output += SveIndexIIOpDeclare.subst(iop)
elif fmt == IndexFormat.ImmReg:
header_output += SveIndexIROpDeclare.subst(iop)
elif fmt == IndexFormat.RegImm:
header_output += SveIndexRIOpDeclare.subst(iop)
elif fmt == IndexFormat.RegReg:
header_output += SveIndexRROpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in ['int8_t', 'int16_t', 'int32_t', 'int64_t']:
substDict = {'targs': type, 'class_name': 'SveIndex'+fmt}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for widening unary SVE instructions
# (always constructive)
def sveWidenUnaryInst(name, Name, opClass, types, op,
predType=PredType.NONE, decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<DElement>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {
SElement srcElem1 = AA64FpOp1_xd[i];
DElement destElem = 0;'''
if predType != PredType.NONE:
code += '''
if (GpOp_xd[i]) {
%(op)s
} else {
destElem = %(dest_elem)s;
}''' % {'op': op,
'dest_elem': 'AA64FpDestMerge_xd[i]'
if predType == PredType.MERGE
else '0'}
else:
code += '''
%(op)s''' % {'op': op}
code += '''
AA64FpDest_xd[i] = destElem;
}'''
iop = ArmInstObjParams(name, 'Sve' + Name,
'SveUnaryPredOp' if predType != PredType.NONE
else 'SveUnaryUnpredOp',
{'code': code, 'op_class': opClass}, [])
if predType != PredType.NONE:
header_output += SveWideningUnaryPredOpDeclare.subst(iop)
else:
header_output += SveWideningUnaryUnpredOpDeclare.subst(iop)
exec_output += SveWideningOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for unary SVE instructions (always constructive)
def sveUnaryInst(name, Name, opClass, types, op, predType=PredType.NONE,
srcRegType=SrcRegType.Vector, decoder='Generic'):
global header_output, exec_output, decoders
op1 = ('AA64FpOp1_x[i]' if srcRegType == SrcRegType.Vector
else 'XOp1' if srcRegType == SrcRegType.Scalar
else 'AA64FpOp1_x[0]')
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {
Element srcElem1 = %s;
Element destElem = 0;''' % op1
if predType != PredType.NONE:
code += '''
if (GpOp_x[i]) {
%(op)s
} else {
destElem = %(dest_elem)s;
}''' % {'op': op,
'dest_elem': 'AA64FpDestMerge_x[i]'
if predType == PredType.MERGE
else '0'}
else:
code += '''
%(op)s''' % {'op': op}
code += '''
AA64FpDest_x[i] = destElem;
}'''
iop = ArmInstObjParams(name, 'Sve' + Name,
'SveUnaryPredOp' if predType != PredType.NONE
else 'SveUnaryUnpredOp',
{'code': code, 'op_class': opClass}, [])
if predType != PredType.NONE:
header_output += SveUnaryPredOpDeclare.subst(iop)
else:
header_output += SveUnaryUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for SVE floating-point conversions (always
# unary, constructive, merging
def sveCvtInst(name, Name, opClass, types, op, direction=CvtDir.Narrow,
decoder='Generic', signed=False):
global header_output, exec_output, decoders
if signed:
mask = "SElement msk = mask(sizeof(DElement)*8);"
assign_code = '''
int sign_bit = bits(destElem, sizeof(DElement)*8 -1);
AA64FpDest_x%(bigElemSuffix)s[i] =
sign_bit? (destElem|~msk): destElem;
''' % {
'bigElemSuffix': 's' if direction == CvtDir.Narrow else 'd'
}
else:
mask = "";
assign_code = '''
AA64FpDest_x%(bigElemSuffix)s[i] = destElem;
''' % {
'bigElemSuffix': 's' if direction == CvtDir.Narrow else 'd'
}
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<%(bigElemType)s>(
xc->tcBase());
%(mask)s
for (unsigned i = 0; i < eCount; i++) {
SElement srcElem1 = AA64FpOp1_x%(bigElemSuffix)s[i] &
mask(sizeof(SElement) * 8);
DElement destElem = 0;
if (GpOp_x%(bigElemSuffix)s[i]) {
%(op)s
%(assign)s;
} else {
AA64FpDest_x%(bigElemSuffix)s[i] =
AA64FpDestMerge_x%(bigElemSuffix)s[i];
}
}
''' % {'bigElemType': 'SElement' if direction == CvtDir.Narrow
else 'DElement',
'op': op, 'mask': mask,
'bigElemSuffix': 's' if direction == CvtDir.Narrow else 'd',
'assign': assign_code
}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveUnaryPredOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveWideningUnaryPredOpDeclare.subst(iop)
exec_output += SveWideningOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for associative SVE reductions
def sveAssocReducInst(name, Name, opClass, types, op, identity,
decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecRegContainer tmpVecC;
auto auxOp1 = tmpVecC.as<Element>();
for (unsigned i = 0; i < eCount; ++i) {
auxOp1[i] = AA64FpOp1_x[i];
}
Element destElem = %(identity)s;
for (unsigned i = 0; i < eCount; i++) {
AA64FpDest_x[i] = 0; // zero upper part
if (GpOp_x[i]) {
const Element& srcElem1 = auxOp1[i];
%(op)s
}
}
AA64FpDest_x[0] = destElem;
''' % {'op': op, 'identity': identity}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveReducOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveReducOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for widening associative SVE reductions
def sveWideningAssocReducInst(name, Name, opClass, types, op, identity,
decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<SElement>(
xc->tcBase());
unsigned eWideCount = ArmStaticInst::getCurSveVecLen<DElement>(
xc->tcBase());
DElement destElem = %(identity)s;
for (unsigned i = 0; i < eCount; i++) {
if (GpOp_xs[i]) {
DElement srcElem1 = AA64FpOp1_xs[i];
%(op)s
}
}
AA64FpDest_xd[0] = destElem;
for (int i = 1; i < eWideCount; i++) {
AA64FpDest_xd[i] = 0;
}
''' % {'op': op, 'identity': identity}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveReducOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveWideningReducOpDeclare.subst(iop)
exec_output += SveWideningOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for non-associative SVE reductions
def sveNonAssocReducInst(name, Name, opClass, types, op, identity,
decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecRegContainer tmpVecC;
auto tmpVec = tmpVecC.as<Element>();
int ePow2Count = 1;
while (ePow2Count < eCount) {
ePow2Count *= 2;
}
for (unsigned i = 0; i < ePow2Count; i++) {
if (i < eCount && GpOp_x[i]) {
tmpVec[i] = AA64FpOp1_x[i];
} else {
tmpVec[i] = %(identity)s;
}
}
unsigned n = ePow2Count;
while (n > 1) {
unsigned max = n;
n = 0;
for (unsigned i = 0; i < max; i += 2) {
Element srcElem1 = tmpVec[i];
Element srcElem2 = tmpVec[i + 1];
Element destElem = 0;
%(op)s
tmpVec[n] = destElem;
n++;
}
}
AA64FpDest_x[0] = tmpVec[0];
for (unsigned i = 1; i < eCount; i++) {
AA64FpDest_x[i] = 0; // zero upper part
}
''' % {'op': op, 'identity': identity}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveReducOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveReducOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for binary SVE instructions with immediate operand
def sveBinImmInst(name, Name, opClass, types, op, predType=PredType.NONE,
decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {'''
if predType != PredType.NONE:
code += '''
const Element& srcElem1 = %s;''' % (
'AA64FpDestMerge_x[i]' if predType == PredType.MERGE else '0')
else:
code += '''
const Element& srcElem1 = AA64FpOp1_x[i];'''
code += '''
Element srcElem2 = imm;
Element destElem = 0;'''
if predType != PredType.NONE:
code += '''
if (GpOp_x[i]) {
%(op)s
} else {
destElem = %(dest_elem)s;
}''' % {'op': op,
'dest_elem': 'AA64FpDestMerge_x[i]'
if predType == PredType.MERGE else '0'}
else:
code += '''
%(op)s''' % {'op': op}
code += '''
AA64FpDest_x[i] = destElem;
}'''
iop = ArmInstObjParams(name, 'Sve' + Name,
'SveBinImmPredOp' if predType != PredType.NONE
else 'SveBinImmUnpredConstrOp',
{'code': code, 'op_class': opClass}, [])
if predType != PredType.NONE:
header_output += SveBinImmPredOpDeclare.subst(iop)
else:
header_output += SveBinImmUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for unary and binary SVE instructions with wide
# immediate operand
def sveWideImmInst(name, Name, opClass, types, op, predType=PredType.NONE,
isUnary=False, decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {'''
# TODO: handle unsigned-to-signed conversion properly...
if isUnary:
code += '''
Element srcElem1 = imm;'''
else:
code += '''
const Element& srcElem1 = AA64FpDestMerge_x[i];
Element srcElem2 = imm;'''
code += '''
Element destElem = 0;'''
if predType != PredType.NONE:
code += '''
if (GpOp_x[i]) {
%(op)s
} else {
destElem = %(dest_elem)s;
}''' % {'op': op,
'dest_elem': 'AA64FpDestMerge_x[i]'
if predType == PredType.MERGE else '0'}
else:
code += '''
%(op)s''' % {'op': op}
code += '''
AA64FpDest_x[i] = destElem;
}'''
iop = ArmInstObjParams(name, 'Sve' + Name,
'Sve%sWideImm%sOp' % (
'Unary' if isUnary else 'Bin',
'Unpred' if predType == PredType.NONE else 'Pred'),
{'code': code, 'op_class': opClass}, [])
if predType == PredType.NONE:
header_output += SveWideImmUnpredOpDeclare.subst(iop)
else:
header_output += SveWideImmPredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for shift SVE instructions with wide elements
def sveShiftByWideElemsInst(name, Name, opClass, types, op,
predType=PredType.NONE, decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecRegContainer tmpVecC;
auto auxOp2 = tmpVecC.as<Element>();
for (unsigned i = 0; i < eCount; i++) {
auxOp2[i] = AA64FpOp2_ud[i];
}
for (unsigned i = 0; i < eCount; i++) {'''
if predType != PredType.NONE:
code += '''
const Element& srcElem1 = AA64FpDestMerge_x[i];'''
else:
code += '''
const Element& srcElem1 = AA64FpOp1_x[i];'''
code += '''
const auto& srcElem2 = auxOp2[
(i * sizeof(Element) * 8) / 64];
Element destElem = 0;'''
if predType != PredType.NONE:
code += '''
if (GpOp_x[i]) {
%(op)s
} else {
destElem = %(dest_elem)s;
}''' % {'op': op,
'dest_elem': 'AA64FpDestMerge_x[i]'
if predType == PredType.MERGE else '0'}
else:
code += '''
%(op)s''' % {'op': op}
code += '''
AA64FpDest_x[i] = destElem;
}'''
iop = ArmInstObjParams(name, 'Sve' + Name,
'SveBinDestrPredOp' if predType != PredType.NONE
else 'SveBinUnpredOp',
{'code': code, 'op_class': opClass}, [])
if predType != PredType.NONE:
header_output += SveBinDestrPredOpDeclare.subst(iop)
else:
header_output += SveBinUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for binary indexed SVE instructions
# (always unpredicated)
def sveBinIdxInst(name, Name, opClass, types, op, decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
// Number of elements in a 128 bit segment
constexpr unsigned ePerSegment = 16 / sizeof(Element);
ArmISA::VecRegContainer tmpC;
auto auxDest = tmpC.as<TPElem>();
for (unsigned i = 0; i < eCount; i++) {
const auto segmentBase = i - i %% ePerSegment;
const auto segmentIdx = segmentBase + index;
const Element& srcElem1 = AA64FpOp1_x[i];
const Element& srcElem2 = AA64FpOp2_x[segmentIdx];
Element destElem = 0;
%(op)s
auxDest[i] = destElem;
}
for (unsigned i = 0; i < eCount; i++) {
AA64FpDest_x[i] = auxDest[i];
}''' % {'op':op}
baseClass = 'SveBinIdxUnpredOp'
iop = ArmInstObjParams(name, 'Sve' + Name, baseClass,
{'code': code, 'op_class': opClass}, [])
header_output += SveBinIdxUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for binary SVE instructions
def sveBinInst(name, Name, opClass, types, op, predType=PredType.NONE,
isDestructive=False, customIterCode=None,
decoder='Generic'):
assert not (predType in (PredType.NONE, PredType.SELECT) and
isDestructive)
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());'''
if customIterCode is None:
code += '''
for (unsigned i = 0; i < eCount; i++) {'''
if predType == PredType.MERGE:
code += '''
const Element& srcElem1 = AA64FpDestMerge_x[i];'''
else:
code += '''
const Element& srcElem1 = AA64FpOp1_x[i];'''
code += '''
const Element& srcElem2 = AA64FpOp2_x[i];
Element destElem = 0;'''
if predType != PredType.NONE:
code += '''
if (GpOp_x[i]) {
%(op)s
} else {
destElem = %(dest_elem)s;
}''' % {'op': op,
'dest_elem':
'AA64FpDestMerge_x[i]' if predType == PredType.MERGE
else '0' if predType == PredType.ZERO
else 'srcElem2'}
else:
code += '''
%(op)s''' % {'op': op}
code += '''
AA64FpDest_x[i] = destElem;
}'''
else:
code += customIterCode
if predType == PredType.NONE:
baseClass = 'SveBinUnpredOp'
elif isDestructive:
baseClass = 'SveBinDestrPredOp'
else:
baseClass = 'SveBinConstrPredOp'
iop = ArmInstObjParams(name, 'Sve' + Name, baseClass,
{'code': code, 'op_class': opClass}, [])
if predType == PredType.NONE:
header_output += SveBinUnpredOpDeclare.subst(iop)
elif isDestructive:
header_output += SveBinDestrPredOpDeclare.subst(iop)
else:
header_output += SveBinConstrPredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for predicate logical instructions
def svePredLogicalInst(name, Name, opClass, types, op,
predType=PredType.ZERO, isFlagSetting=False,
decoder='Generic'):
global header_output, exec_output, decoders
assert predType in (PredType.ZERO, PredType.SELECT)
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecPredRegContainer tmpPredC;
auto auxGpOp = tmpPredC.as<Element>();
for (unsigned i = 0; i < eCount; i++) {
auxGpOp[i] = GpOp_x[i];
}
for (unsigned i = 0; i < eCount; i++) {
bool srcElem1 = POp1_x[i];
bool srcElem2 = POp2_x[i];
bool destElem = false;
if (auxGpOp[i]) {
%(op)s
} else {
destElem = %(dest_elem)s;
}
PDest_x[i] = destElem;
}''' % {'op': op,
'dest_elem': 'false' if predType == PredType.ZERO
else 'srcElem2'}
extraPrologCode = ''
if isFlagSetting:
code += '''
CondCodesNZ = (destPred.firstActive(auxGpOp, eCount) << 1) |
destPred.noneActive(auxGpOp, eCount);
CondCodesC = !destPred.lastActive(auxGpOp, eCount);
CondCodesV = 0;'''
extraPrologCode += '''
auto& destPred = PDest;'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SvePredLogicalOp',
{'code': code, 'op_class': opClass}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SvePredLogicalOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for predicate permute instructions
def svePredBinPermInst(name, Name, opClass, types, iterCode,
decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());'''
code += iterCode
iop = ArmInstObjParams(name, 'Sve' + Name, 'SvePredBinPermOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveBinUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for SVE compare instructions
# NOTE: compares are all predicated zeroing
def sveCmpInst(name, Name, opClass, types, op, isImm=False,
decoder='Generic'):
global header_output, exec_output, decoders
extraPrologCode = '''
auto& destPred = PDest;'''
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecPredRegContainer tmpPredC;
auto tmpPred = tmpPredC.as<Element>();
for (unsigned i = 0; i < eCount; ++i)
tmpPred[i] = GpOp_x[i];
destPred.reset();
for (unsigned i = 0; i < eCount; i++) {
const Element& srcElem1 = AA64FpOp1_x[i];
[[maybe_unused]] %(src_elem_2_ty)s srcElem2 = %(src_elem_2)s;
bool destElem = false;
if (tmpPred[i]) {
%(op)s
} else {
destElem = false;
}
PDest_x[i] = destElem;
}''' % {'op': op,
'src_elem_2_ty': 'Element' if isImm else 'const Element&',
'src_elem_2': 'imm' if isImm else 'AA64FpOp2_x[i]'}
iop = ArmInstObjParams(name, 'Sve' + Name,
'SveCmpImmOp' if isImm else 'SveCmpOp',
{'code': code, 'op_class': opClass}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
if isImm:
header_output += SveCmpImmOpDeclare.subst(iop)
else:
header_output += SveCmpOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for ternary SVE intructions (always predicated -
# merging)
def sveTerInst(name, Name, opClass, types, op, decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {
const Element& srcElem1 = AA64FpOp1_x[i];
const Element& srcElem2 = AA64FpOp2_x[i];
Element destElem = AA64FpDestMerge_x[i];
if (GpOp_x[i]) {
%(op)s
}
AA64FpDest_x[i] = destElem;
}''' % {'op': op}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveTerPredOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveTerPredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for ternary SVE instructions with indexed operand
def sveTerIdxInst(name, Name, opClass, types, op, decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
// Number of elements in a 128 bit segment
constexpr unsigned ePerSegment = 16 / sizeof(Element);
ArmISA::VecRegContainer tmpC;
auto auxDest = tmpC.as<TPElem>();
for (unsigned i = 0; i < eCount; i++) {
const auto segmentBase = i - i % ePerSegment;
const auto segmentIdx = segmentBase + index;
const Element& srcElem1 = AA64FpOp1_x[i];
const Element& srcElem2 = AA64FpOp2_x[segmentIdx];
Element destElem = AA64FpDestMerge_x[i];
'''
code += '''
%(op)s
auxDest[i] = destElem;
}
for (unsigned i = 0; i < eCount; i++) {
AA64FpDest_x[i] = auxDest[i];
}''' % {'op': op}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveBinIdxUnpredOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveBinIdxUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for ternary SVE intructions with immediate operand
# (always unpredicated)
def sveTerImmInst(name, Name, opClass, types, op, decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {
const Element& srcElem2 = AA64FpOp2_x[i];
Element srcElem3 = imm;
Element destElem = AA64FpDestMerge_x[i];
%(op)s
AA64FpDest_x[i] = destElem;
}''' % {'op': op}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveTerImmUnpredOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveTerImmUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generates definitions for PTRUE and PTRUES instructions.
def svePtrueInst(name, Name, opClass, types, isFlagSetting=False,
decoder='Generic'):
global header_output, exec_output, decoders
extraPrologCode = '''
auto& destPred = PDest;'''
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
unsigned predCount = sveDecodePredCount(imm, eCount);
destPred.reset();
for (unsigned i = 0; i < eCount; i++) {
PDest_x[i] = (i < predCount);
}'''
if isFlagSetting:
code += '''
CondCodesNZ = (destPred.firstActive(destPred, eCount) << 1) |
destPred.noneActive(destPred, eCount);
CondCodesC = !destPred.lastActive(destPred, eCount);
CondCodesV = 0;'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SvePtrueOp',
{'code': code, 'op_class': opClass}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SvePtrueOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definitions for integer CMP<cc> instructions
def sveIntCmpInst(name, Name, opClass, types, op, wideop = False,
decoder = 'Generic'):
global header_output, exec_output, decoders
signed = 'int8_t' in types
srcType = 'Element'
op2Suffix = 'x'
if wideop:
srcType = 'int64_t' if signed else 'uint64_t'
op2Suffix = 'sd' if signed else 'ud'
extraPrologCode = '''
auto& destPred = PDest;'''
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecPredRegContainer tmpPredC;
auto tmpPred = tmpPredC.as<Element>();
for (unsigned i = 0; i < eCount; ++i)
tmpPred[i] = GpOp_x[i];
destPred.reset();
for (unsigned i = 0; i < eCount; ++i) {
%(srcType)s srcElem1 = (%(srcType)s) AA64FpOp1_x[i];
%(srcType)s srcElem2 = AA64FpOp2_%(op2Suffix)s[%(op2Index)s];
bool destElem = false;
if (tmpPred[i]) {
%(op)s
}
PDest_x[i] = destElem;
}
CondCodesNZ = (destPred.firstActive(tmpPred, eCount) << 1) |
destPred.noneActive(tmpPred, eCount);
CondCodesC = !destPred.lastActive(tmpPred, eCount);
CondCodesV = 0;''' % {
'op': op,
'srcType': srcType,
'op2Suffix': op2Suffix,
'op2Index': '(i * sizeof(Element)) / 8' if wideop else 'i'
}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveIntCmpOp',
{
'code': code,
'op_class': opClass,
'op2IsWide': 'true' if wideop else 'false',
}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SveIntCmpOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definitions for integer CMP<cc> instructions (with immediate)
def sveIntCmpImmInst(name, Name, opClass, types, op, decoder = 'Generic'):
global header_output, exec_output, decoders
extraPrologCode = '''
auto& destPred = PDest;'''
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecPredRegContainer tmpPredC;
auto tmpPred = tmpPredC.as<Element>();
for (unsigned i = 0; i < eCount; ++i)
tmpPred[i] = GpOp_x[i];
destPred.reset();
for (unsigned i = 0; i < eCount; ++i) {
Element srcElem1 = AA64FpOp1_x[i];
Element srcElem2 = static_cast<Element>(imm);
bool destElem = false;
if (tmpPred[i]) {
%(op)s
}
PDest_x[i] = destElem;
}
CondCodesNZ = (destPred.firstActive(tmpPred, eCount) << 1) |
destPred.noneActive(tmpPred, eCount);
CondCodesC = !destPred.lastActive(tmpPred, eCount);
CondCodesV = 0;'''%{'op': op}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveIntCmpImmOp',
{'code': code, 'op_class': opClass,}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SveIntCmpImmOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definitions for SVE element count instructions
def sveElemCountInst(name, Name, opClass, types, op,
destType = DestType.Scalar, dstIs32b = False,
dstAcc = True, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
unsigned count = sveDecodePredCount(pattern, eCount);
'''
if destType == DestType.Vector:
code += '''
for (unsigned i = 0; i < eCount; ++i) {
Element srcElem1 = AA64FpDestMerge_x[i];
Element destElem = 0;
%(op)s
AA64FpDest_x[i] = destElem;
}'''%{'op': op}
else:
if 'uint16_t' in types:
if dstIs32b:
dstType = 'uint32_t'
else:
dstType = 'uint64_t'
else:
if dstIs32b:
dstType = 'int32_t'
else:
dstType = 'int64_t'
if dstAcc:
code += '''
%(dstType)s srcElem1 = XDest;
'''%{'dstType': dstType}
code += '''
%(dstType)s destElem = 0;
%(op)s;
XDest = destElem;
'''%{'op': op, 'dstType': dstType}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveElemCountOp',
{'code': code, 'op_class': opClass, 'dstIsVec': destType,
'dstIs32b': 'true' if dstIs32b else 'false'}, [])
header_output += SveElemCountOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict);
def svePartBrkInst(name, Name, opClass, isFlagSetting, predType, whenBrk,
decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<uint8_t>(
xc->tcBase());
bool dobreak = false;
ArmISA::VecPredRegContainer tmpPredC;
auto auxGpOp = tmpPredC.as<uint8_t>();
for (unsigned i = 0; i < eCount; ++i) {
auxGpOp[i] = GpOp_ub[i];
}
for (unsigned i = 0; i < eCount; ++i) {
bool element = POp1_ub[i] == 1;
if (auxGpOp[i]) {'''
breakCode = '''
dobreak = dobreak || element;'''
if whenBrk == Break.Before:
code += breakCode
code += '''
PDest_ub[i] = !dobreak;'''
if whenBrk == Break.After:
code += breakCode
code += '''
}'''
if predType == PredType.ZERO:
code += ''' else {
PDest_ub[i] = 0;
}'''
elif predType == PredType.MERGE:
code += ''' else {
PDest_ub[i] = PDestMerge_ub[i];
}'''
code += '''
}'''
extraPrologCode = ''
if isFlagSetting:
code += '''
CondCodesNZ = (destPred.firstActive(auxGpOp, eCount) << 1) |
destPred.noneActive(auxGpOp, eCount);
CondCodesC = !destPred.lastActive(auxGpOp, eCount);
CondCodesV = 0;'''
extraPrologCode += '''
auto& destPred = PDest;'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SvePartBrkOp',
{'code': code, 'op_class': opClass,
'isMerging':
'true' if predType == PredType.MERGE
else 'false'}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SvePartBrkOpDeclare.subst(iop)
exec_output += SveNonTemplatedOpExecute.subst(iop)
def svePartBrkPropPrevInst(name, Name, opClass, isFlagSetting, whenBrk,
decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<uint8_t>(
xc->tcBase());
bool last = POp1_ub.lastActive(GpOp_ub, eCount);
ArmISA::VecPredRegContainer tmpPredC;
auto auxGpOp = tmpPredC.as<uint8_t>();
for (unsigned i = 0; i < eCount; ++i) {
auxGpOp[i] = GpOp_ub[i];
}
for (unsigned i = 0; i < eCount; ++i) {
if (auxGpOp[i]) {'''
breakCode = '''
last = last && (POp2_ub[i] == 0);'''
if whenBrk == Break.Before:
code += breakCode
code += '''
PDest_ub[i] = last;'''
if whenBrk == Break.After:
code += breakCode
code += '''
} else {
PDest_ub[i] = 0;
}
}'''
extraPrologCode = ''
if isFlagSetting:
code += '''
CondCodesNZ = (destPred.firstActive(auxGpOp, eCount) << 1) |
destPred.noneActive(auxGpOp, eCount);
CondCodesC = !destPred.lastActive(auxGpOp, eCount);
CondCodesV = 0;'''
extraPrologCode += '''
auto& destPred = PDest;'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SvePartBrkPropOp',
{'code': code, 'op_class': opClass}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SvePartBrkPropOpDeclare.subst(iop)
exec_output += SveNonTemplatedOpExecute.subst(iop)
def svePartBrkPropNextInst(name, Name, opClass, isFlagSetting,
decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<uint8_t>(
xc->tcBase());
bool last = POp1_ub.lastActive(GpOp_ub, eCount);
for (unsigned i = 0; i < eCount; i++) {
if (!last) {
PDest_ub[i] = 0;
} else {
PDest_ub[i] = PDestMerge_ub[i];
}
}'''
extraPrologCode = ''
if isFlagSetting:
code += '''
VecPredRegT<uint8_t, MaxSveVecLenInBytes, false, false>::Container c;
VecPredRegT<uint8_t, MaxSveVecLenInBytes, false, false> predOnes(c);
for (unsigned i = 0; i < eCount; i++) {
predOnes[i] = 1;
}
CondCodesNZ = (destPred.firstActive(predOnes, eCount) << 1) |
destPred.noneActive(predOnes, eCount);
CondCodesC = !destPred.lastActive(predOnes, eCount);
CondCodesV = 0;'''
extraPrologCode += '''
auto& destPred = PDest;'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SvePartBrkPropOp',
{'code': code, 'op_class': opClass}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SvePartBrkPropOpDeclare.subst(iop)
exec_output += SveNonTemplatedOpExecute.subst(iop)
# Generate definitions for scalar select instructions
def sveSelectInst(name, Name, opClass, types, op, isCond,
destType = DstRegType.Scalar, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
int last;
for (last = eCount - 1; last >= 0; --last) {
if (GpOp_x[last]) {
break;
}
}
'''
if isCond:
code += '''
if (last >= 0) {'''
code += '''
Element destElem;
%(op)s'''%{'op': op}
if destType == DstRegType.Vector:
code += '''
for (unsigned i = 0; i < eCount; ++i)
AA64FpDest_x[i] = destElem;'''
elif destType == DstRegType.Scalar:
code += '''
XDest = destElem;'''
elif destType == DstRegType.SimdFpScalar:
code += '''
AA64FpDest_x[0] = destElem;'''
if isCond:
code += '''
}'''
if destType == DstRegType.Scalar:
code += ''' else {
XDest = (Element) XDest;
}'''
elif destType == DstRegType.Vector:
code += ''' else {
for (unsigned i = 0; i < eCount; ++i)
AA64FpDest_x[i] = AA64FpDestMerge_x[i];
}'''
elif destType == DstRegType.SimdFpScalar:
code += ''' else {
AA64FpDest_x[0] = AA64FpDestMerge_x[0];
}
'''
if destType == DstRegType.SimdFpScalar:
# This section will extend zeros to the simdFP scalar
# intructions for lasta/b and Clasta/b
code += '''
for (int i = 1; i < eCount; ++i) {
AA64FpDest_x[i] = (Element)0x0;
}
'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveSelectOp',
{'code': code, 'op_class': opClass,
'isCond': 'true' if isCond else 'false',
'isScalar': 'true'
if destType == DstRegType.Scalar else 'false',
'isSimdFp': 'true'
if destType == DstRegType.SimdFpScalar
else 'false'},
[])
header_output += SveSelectOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definitions for PNEXT (find next active predicate)
# instructions
def svePNextInst(name, Name, opClass, types, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecPredRegContainer tmpPredC;
auto auxGpOp = tmpPredC.as<Element>();
for (unsigned i = 0; i < eCount; ++i) {
auxGpOp[i] = GpOp_x[i];
}
int last;
for (last = eCount - 1; last >= 0; --last) {
if (POp1_x[last]) {
break;
}
}
int next = last + 1;
while (next < eCount && GpOp_x[next] == 0) {
next++;
}
destPred.reset();
if (next < eCount) {
PDest_x[next] = 1;
}
CondCodesNZ = (destPred.firstActive(auxGpOp, eCount) << 1) |
destPred.noneActive(auxGpOp, eCount);
CondCodesC = !destPred.lastActive(auxGpOp, eCount);
CondCodesV = 0;'''
extraPrologCode = '''
auto& destPred = PDest;'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveUnaryPredPredOp',
{'code': code, 'op_class': opClass}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SveUnaryPredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definitions for PFIRST (set first active predicate)
# instructions
def svePFirstInst(name, Name, opClass, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecPredRegContainer tmpPredC;
auto auxGpOp = tmpPredC.as<Element>();
for (unsigned i = 0; i < eCount; ++i)
auxGpOp[i] = GpOp_x[i];
int first = -1;
for (int i = 0; i < eCount; ++i) {
if (auxGpOp[i] && first == -1) {
first = i;
}
}
for (int i = 0; i < eCount; ++i) {
PDest_x[i] = PDestMerge_x[i];
}
if (first >= 0) {
PDest_x[first] = 1;
}
CondCodesNZ = (destPred.firstActive(auxGpOp, eCount) << 1) |
destPred.noneActive(auxGpOp, eCount);
CondCodesC = !destPred.lastActive(auxGpOp, eCount);
CondCodesV = 0;'''
extraPrologCode = '''
auto& destPred = PDest;'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveUnaryPredPredOp',
{'code': code, 'op_class': opClass}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SveUnaryPredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
substDict = {'targs' : 'uint8_t',
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definitions for SVE TBL instructions
def sveTblInst(name, Name, opClass, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
for (int i = 0; i < eCount; ++i) {
Element idx = AA64FpOp2_x[i];
Element val;
if (idx < eCount) {
val = AA64FpOp1_x[idx];
} else {
val = 0;
}
AA64FpDest_x[i] = val;
}'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveTblOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveBinUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in unsignedTypes:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definitions for SVE Unpack instructions
def sveUnpackInst(name, Name, opClass, sdtypes, unpackHalf,
regType, decoder = 'Generic'):
global header_output, exec_output, decoders
extraPrologCode = '''
auto& destPred = PDest;'''
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<DElement>(
xc->tcBase());'''
if unpackHalf == Unpack.Low:
if regType == SrcRegType.Predicate:
code += '''
ArmISA::VecPredRegContainer tmpPredC;
auto auxPOp1 = tmpPredC.as<SElement>();
for (int i = 0; i < eCount; ++i) {
auxPOp1[i] = POp1_xs[i];
}'''
else:
code += '''
ArmISA::VecRegContainer tmpVecC;
auto auxOp1 = tmpVecC.as<SElement>();
for (int i = 0; i < eCount; ++i) {
auxOp1[i] = AA64FpOp1_xs[i];
}'''
code += '''
for (int i = 0; i < eCount; ++i) {'''
if regType == SrcRegType.Predicate:
if unpackHalf == Unpack.High:
code +='''
const SElement& srcElem1 = POp1_xs[i + eCount];'''
else:
code +='''
const SElement& srcElem1 = auxPOp1[i];'''
code += '''
destPred.setRaw(i, 0);
PDest_xd[i] = srcElem1;'''
else:
if unpackHalf == Unpack.High:
code +='''
const SElement& srcElem1 = AA64FpOp1_xs[i + eCount];'''
else:
code +='''
const SElement& srcElem1 = auxOp1[i];'''
code += '''
AA64FpDest_xd[i] = static_cast<DElement>(srcElem1);'''
code += '''
}
'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveUnpackOp',
{'code': code, 'op_class': opClass}, [])
if regType == SrcRegType.Predicate:
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SveUnpackOpDeclare.subst(iop)
exec_output += SveWideningOpExecute.subst(iop)
for srcType, dstType in sdtypes:
substDict = {'targs': srcType + ', ' + dstType,
'class_name': 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definition for SVE predicate test instructions
def svePredTestInst(name, Name, opClass, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<uint8_t>(
xc->tcBase());
CondCodesNZ = (POp1_ub.firstActive(GpOp_ub, eCount) << 1) |
POp1_ub.noneActive(GpOp_ub, eCount);
CondCodesC = !POp1_ub.lastActive(GpOp_ub, eCount);
CondCodesV = 0;'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SvePredTestOp',
{'code': code, 'op_class': opClass}, [])
header_output += SvePredicateTestOpDeclare.subst(iop)
exec_output += SveNonTemplatedOpExecute.subst(iop)
# Generate definition for SVE predicate compact operations
def sveCompactInst(name, Name, opClass, types, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecRegContainer tmpVecC;
auto auxOp1 = tmpVecC.as<Element>();
for (unsigned i = 0; i < eCount; ++i) {
auxOp1[i] = AA64FpOp1_x[i];
}
unsigned x = 0;
for (unsigned i = 0; i < eCount; ++i) {
AA64FpDest_x[i] = 0;
if (GpOp_x[i]) {
AA64FpDest_x[x] = auxOp1[i];
x++;
}
}'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveUnaryPredOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveUnaryPredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs': type, 'class_name': 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definition for unary SVE predicate instructions with implicit
# source operand (PFALSE, RDFFR(S))
def svePredUnaryWImplicitSrcInst(name, Name, opClass, op,
predType=PredType.NONE, isFlagSetting=False, decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + op
if isFlagSetting:
code += '''
CondCodesNZ = (destPred.firstActive(GpOp, eCount) << 1) |
destPred.noneActive(GpOp, eCount);
CondCodesC = !destPred.lastActive(GpOp, eCount);
CondCodesV = 0;'''
extraPrologCode = '''
[[maybe_unused]] auto& destPred = PDest;'''
baseClass = ('SvePredUnaryWImplicitSrcOp' if predType == PredType.NONE
else 'SvePredUnaryWImplicitSrcPredOp')
iop = ArmInstObjParams(name, 'Sve' + Name, baseClass,
{'code': code, 'op_class': opClass}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
if predType == PredType.NONE:
header_output += SvePredUnaryOpWImplicitSrcDeclare.subst(iop)
else:
header_output += SvePredUnaryPredOpWImplicitSrcDeclare.subst(iop)
exec_output += SveNonTemplatedOpExecute.subst(iop)
# Generate definition for SVE instructions writing to the FFR (SETFFR,
# WRFFR)
def svePredWriteFfrInst(name, Name, opClass, op, isSetFfr,
decoder='Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + op
extraPrologCode = '''
[[maybe_unused]] auto& destPred = Ffr;'''
baseClass = ('SveWImplicitSrcDstOp' if isSetFfr
else 'SvePredUnaryWImplicitDstOp')
iop = ArmInstObjParams(name, 'Sve' + Name, baseClass,
{'code': code, 'op_class': opClass}, [])
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
if isSetFfr:
header_output += SveOpWImplicitSrcDstDeclare.subst(iop)
else:
header_output += SvePredUnaryOpWImplicitDstDeclare.subst(iop)
exec_output += SveNonTemplatedOpExecute.subst(iop)
# Generate definition for SVE Ext instruction
def sveExtInst(name, Name, opClass, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecRegContainer tmpVecC;
auto auxOp1 = tmpVecC.as<Element>();
for (unsigned i = 0; i < eCount; ++i) {
auxOp1[i] = AA64FpOp1_x[i];
}
uint64_t pos = imm;
if (pos >= eCount)
pos = 0;
for (int i = 0; i < eCount; ++i, ++pos)
{
if (pos < eCount)
AA64FpDest_x[i] = AA64FpDestMerge_x[pos];
else
AA64FpDest_x[i] = auxOp1[pos-eCount];
}
'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveBinImmUnpredDestrOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveBinImmUnpredOpDeclare.subst(iop);
exec_output += SveOpExecute.subst(iop)
substDict = {'targs': 'uint8_t', 'class_name': 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definition for SVE Slice instruction
def sveSpliceInst(name, Name, opClass, types, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
ArmISA::VecRegContainer tmpVecC;
auto auxDest = tmpVecC.as<Element>();
int firstelem = -1, lastelem = -2;
for (int i = 0; i < eCount; ++i) {
if (GpOp_x[i]) {
lastelem = i;
if (firstelem < 0)
firstelem = i;
}
}
int x = 0;
for (int i = firstelem; i <= lastelem; ++i, ++x) {
auxDest[x] = AA64FpDestMerge_x[i];
}
int remaining = eCount - x;
for (int i = 0; i < remaining; ++i, ++x) {
auxDest[x] = AA64FpOp2_x[i];
}
for (int i = 0; i < eCount; ++i) {
AA64FpDest_x[i] = auxDest[i];
}
'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveBinDestrPredOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveBinDestrPredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs': type, 'class_name': 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definition for SVE DUP (index) instruction
def sveDupIndexInst(name, Name, opClass, types, decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
Element srcElem1 = 0;
if (imm < eCount) {
srcElem1 = AA64FpOp1_x[imm];
}
for (int i = 0; i < eCount; ++i) {
AA64FpDest_x[i] = srcElem1;
}'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveBinImmIdxUnpredOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveBinImmUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs': type, 'class_name': 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definition for SVE reverse elements instructions
def sveReverseElementsInst(name, Name, opClass, types,
srcType = SrcRegType.Vector, decoder = 'Generic'):
assert srcType in (SrcRegType.Vector, SrcRegType.Predicate)
global header_output, exec_output, decoders
extraPrologCode = '''
auto& destPred = PDest;'''
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());'''
if srcType == SrcRegType.Predicate:
code += '''
ArmISA::VecPredRegContainer tmpPredC;
auto auxPOp1 = tmpPredC.as<Element>();
for (unsigned i = 0; i < eCount; ++i) {
uint8_t v = POp1_x.getRaw(i);
auxPOp1.setRaw(i, v);
}
PDest_x[0] = 0;'''
else:
code += '''
ArmISA::VecRegContainer tmpRegC;
auto auxOp1 = tmpRegC.as<Element>();
for (unsigned i = 0; i < eCount; ++i) {
auxOp1[i] = AA64FpOp1_x[i];
}'''
code += '''
for (int i = 0; i < eCount; ++i) {'''
if srcType == SrcRegType.Vector:
code += '''
AA64FpDest_x[i] = auxOp1[eCount - i - 1];'''
else:
code += '''
destPred.setRaw(i, auxPOp1.getRaw(eCount - i - 1));'''
code += '''
}'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveUnaryUnpredOp',
{'code': code, 'op_class': opClass}, [])
if srcType == SrcRegType.Predicate:
iop.snippets['code'] = extraPrologCode + iop.snippets['code']
header_output += SveUnaryUnpredOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs': type, 'class_name': 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definition for shift & insert instructions
def sveShiftAndInsertInst(name, Name, opClass, types,
srcType = SrcRegType.Scalar, decoder = 'Generic'):
assert srcType in (SrcRegType.SimdFpScalar, SrcRegType.Scalar)
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());'''
if srcType == SrcRegType.Scalar:
code += '''
auto& srcElem1 = XOp1;'''
elif srcType == SrcRegType.SimdFpScalar:
code += '''
auto& srcElem1 = AA64FpOp1_x[0];'''
code += '''
for (int i = eCount - 1; i > 0; --i) {
AA64FpDest_x[i] = AA64FpDestMerge_x[i-1];
}
AA64FpDest_x[0] = srcElem1;'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveUnarySca2VecUnpredOp',
{'code': code, 'op_class': opClass,
'isSimdFp': 'true' if srcType == SrcRegType.SimdFpScalar
else 'false'}, [])
header_output += SveShiftAndInsertOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs': type, 'class_name': 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definition for DOT instructions
def sveDotInst(name, Name, opClass, types, isIndexed = True):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());
for (int i = 0; i < eCount; ++i) {'''
if isIndexed:
code += '''
int segbase = i - i % (16 / sizeof(Element));
int s = segbase + imm;'''
code += '''
DElement res = AA64FpDest_xd[i];
DElement srcElem1, srcElem2;
for (int j = 0; j <= 3; ++j) {
srcElem1 = static_cast<DElement>(AA64FpOp1_xs[4 * i + j]);'''
if isIndexed:
code += '''
srcElem2 = static_cast<DElement>(AA64FpOp2_xs[4 * s + j]);'''
else:
code += '''
srcElem2 = static_cast<DElement>(AA64FpOp2_xs[4 * i + j]);'''
code += '''
res += srcElem1 * srcElem2;
}
AA64FpDestMerge_xd[i] = res;
}'''
iop = ArmInstObjParams(name, 'Sve' + Name,
'SveDotProdIdxOp' if isIndexed else
'SveDotProdOp',
{'code': code, 'op_class': opClass}, [])
if isIndexed:
header_output += SveWideningTerImmOpDeclare.subst(iop)
else:
header_output += SveWideningTerOpDeclare.subst(iop)
exec_output += SveWideningOpExecute.subst(iop)
for type in types:
substDict = {'targs': type, 'class_name': 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definition for ordered reduction
def sveOrderedReduction(name, Name, opClass, types, op,
decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());'''
code += '''
Element destElem = AA64FpDestMerge_x[0];
for (int i = 0; i < eCount; ++i) {
if (GpOp_x[i]) {
Element srcElem1 = AA64FpOp1_x[i];
%(op)s
}
}
for (int i = 1; i < eCount; ++i) {
AA64FpDest_x[i] = 0;
}
AA64FpDest_x[0] = destElem;'''%{'op': op}
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveOrdReducOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveReducOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definitions for complex addition instructions
def sveComplexAddInst(name, Name, opClass, types,
decoder = 'Generic'):
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());'''
code += '''
bool sub_i = (rot == 1);
bool sub_r = (rot == 3);
for (int i = 0; i < eCount / 2; ++i) {
Element acc_r = AA64FpOp1_x[2 * i];
Element acc_i = AA64FpOp1_x[2 * i + 1];
Element elt2_r = AA64FpOp2_x[2 * i];
Element elt2_i = AA64FpOp2_x[2 * i + 1];
FPSCR fpscr;
if (GpOp_x[2 * i]) {
if (sub_i) {
elt2_i = fplibNeg<Element>(elt2_i);
}
fpscr = FpscrExc & ~FpscrAhpMask;
acc_r = fplibAdd<Element>(acc_r, elt2_i, fpscr);
FpscrExc = fpscr;
}
if (GpOp_x[2 * i + 1]) {
if (sub_r) {
elt2_r = fplibNeg<Element>(elt2_r);
}
fpscr = FpscrExc & ~FpscrAhpMask;
acc_i = fplibAdd<Element>(acc_i, elt2_r, fpscr);
FpscrExc = fpscr;
}
AA64FpDest_x[2 * i] = acc_r;
AA64FpDest_x[2 * i + 1] = acc_i;
}
'''
iop = ArmInstObjParams(name, 'Sve' + Name, 'SveComplexOp',
{'code': code, 'op_class': opClass}, [])
header_output += SveComplexOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
# Generate definitions for complex multiply and accumulate instructions
def sveComplexMulAddInst(name, Name, opClass, types,
predType=PredType.NONE, decoder='Generic'):
assert predType in (PredType.NONE, PredType.MERGE)
global header_output, exec_output, decoders
code = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<Element>(
xc->tcBase());'''
code += '''
uint32_t sel_a = bits(rot, 0);
uint32_t sel_b = sel_a ? 0 : 1;
bool neg_i = bits(rot, 1);
bool neg_r = bits(rot, 0) != bits(rot, 1);'''
if predType == PredType.NONE:
code += '''
uint32_t eltspersegment = 16 / (2 * sizeof(Element));'''
code += '''
ArmISA::VecRegContainer tmpC;
auto auxDest = tmpC.as<TPElem>();
for (int i = 0; i < eCount / 2; ++i) {'''
if predType == PredType.NONE:
code += '''
uint32_t segmentbase = i - (i % eltspersegment);
uint32_t s = segmentbase + imm;'''
else:
code += '''
uint32_t s = i;'''
code += '''
Element addend_r = AA64FpDestMerge_x[2 * i];
Element addend_i = AA64FpDestMerge_x[2 * i + 1];
Element elt1_a = AA64FpOp1_x[2 * i + sel_a];
Element elt2_a = AA64FpOp2_x[2 * s + sel_a];
Element elt2_b = AA64FpOp2_x[2 * s + sel_b];
FPSCR fpscr;
'''
if predType != PredType.NONE:
code += '''
if (GpOp_x[2 * i]) {'''
code += '''
if (neg_r) {
elt2_a = fplibNeg<Element>(elt2_a);
}
fpscr = FpscrExc & ~FpscrAhpMask;
addend_r = fplibMulAdd<Element>(addend_r, elt1_a, elt2_a, fpscr);
FpscrExc = fpscr;'''
if predType != PredType.NONE:
code += '''
}'''
if predType != PredType.NONE:
code += '''
if (GpOp_x[2 * i + 1]) {'''
code += '''
if (neg_i) {
elt2_b = fplibNeg<Element>(elt2_b);
}
fpscr = FpscrExc & ~FpscrAhpMask;
addend_i = fplibMulAdd<Element>(addend_i, elt1_a, elt2_b, fpscr);
FpscrExc = fpscr;'''
if predType != PredType.NONE:
code += '''
}'''
code += '''
auxDest[2 * i] = addend_r;
auxDest[2 * i + 1] = addend_i;
}
for (unsigned i = 0; i < eCount; i++) {
AA64FpDest_x[i] = auxDest[i];
}
'''
iop = ArmInstObjParams(name, 'Sve' + Name,
'SveComplexIdxOp' if predType == PredType.NONE
else 'SveComplexOp',
{'code': code, 'op_class': opClass}, [])
if predType == PredType.NONE:
header_output += SveComplexIndexOpDeclare.subst(iop)
else:
header_output += SveComplexOpDeclare.subst(iop)
exec_output += SveOpExecute.subst(iop)
for type in types:
substDict = {'targs' : type,
'class_name' : 'Sve' + Name}
exec_output += SveOpExecDeclare.subst(substDict)
fpTypes = ('uint16_t', 'uint32_t', 'uint64_t')
signedTypes = ('int8_t', 'int16_t', 'int32_t', 'int64_t')
unsignedTypes = ('uint8_t', 'uint16_t', 'uint32_t', 'uint64_t')
smallSignedTypes = ('int8_t', 'int16_t', 'int32_t')
bigSignedTypes = ('int16_t', 'int32_t', 'int64_t')
smallUnsignedTypes = ('uint8_t', 'uint16_t', 'uint32_t')
bigUnsignedTypes = ('uint16_t', 'uint32_t', 'uint64_t')
unsignedWideSDTypes = (('uint8_t', 'uint16_t'),
('uint16_t', 'uint32_t'), ('uint32_t', 'uint64_t'))
signedWideSDTypes = (('int8_t', 'int16_t'),
('int16_t', 'int32_t'), ('int32_t', 'int64_t'))
# ABS
absCode = 'destElem = (Element) std::abs(srcElem1);'
sveUnaryInst('abs', 'Abs', 'SimdAluOp', signedTypes, absCode,
PredType.MERGE)
# ADD (immediate)
sveWideImmInst('add', 'AddImm', 'SimdAddOp', unsignedTypes, addCode, False)
# ADD (vectors, predicated)
addCode = 'destElem = srcElem1 + srcElem2;'
sveBinInst('add', 'AddPred', 'SimdAddOp', unsignedTypes, addCode,
PredType.MERGE, True)
# ADD (vectors, unpredicated)
addCode = 'destElem = srcElem1 + srcElem2;'
sveBinInst('add', 'AddUnpred', 'SimdAddOp', unsignedTypes, addCode)
# ADDPL
addvlCode = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<uint%d_t>(
xc->tcBase());
XDest = XOp1 + eCount * (int64_t) imm;
'''
buildXImmDataInst('addpl', addvlCode % 64, buildCc=False)
# ADDVL
buildXImmDataInst('addvl', addvlCode % 8, buildCc=False)
# ADR
adrCode = '''
if (offsetFormat == SveAdrOffsetUnpackedSigned) {
srcElem2 = sext<32>(srcElem2 & mask(32));
} else if (offsetFormat == SveAdrOffsetUnpackedUnsigned) {
srcElem2 = srcElem2 & mask(32);
}
destElem = srcElem1 + srcElem2 * mult;
'''
sveAdrInst('adr', 'Adr', 'SimdAddOp', ('uint32_t', 'uint64_t'), adrCode)
# AND (immediate)
andCode = 'destElem = srcElem1 & srcElem2;'
sveWideImmInst('and', 'AndImm', 'SimdAluOp', ('uint64_t',), andCode)
# AND (vectors, predicated)
sveBinInst('and', 'AndPred', 'SimdAluOp', unsignedTypes, andCode,
PredType.MERGE, True)
# AND (vectors, unpredicated)
andCode = 'destElem = srcElem1 & srcElem2;'
sveBinInst('and', 'AndUnpred', 'SimdAluOp', ('uint64_t',), andCode)
# AND, ANDS (predicates)
svePredLogicalInst('and', 'PredAnd', 'SimdPredAluOp', ('uint8_t',),
andCode)
svePredLogicalInst('ands', 'PredAnds', 'SimdPredAluOp', ('uint8_t',),
andCode, isFlagSetting=True)
# ANDV
andvCode = 'destElem &= srcElem1;'
sveAssocReducInst('andv', 'Andv', 'SimdReduceAluOp', unsignedTypes,
andvCode, 'std::numeric_limits<Element>::max()')
# ASR (immediate, predicated)
asrCode = '''
int sign_bit = bits(srcElem1, sizeof(Element) * 8 - 1);
if (srcElem2 == 0) {
destElem = srcElem1;
} else if (srcElem2 >= sizeof(Element) * 8) {
destElem = sign_bit ? std::numeric_limits<Element>::max() : 0;
} else {
destElem = srcElem1 >> srcElem2;
if (sign_bit) {
destElem |= ~mask(sizeof(Element) * 8 - srcElem2);
}
}
'''
sveBinImmInst('asr', 'AsrImmPred', 'SimdAluOp', unsignedTypes, asrCode,
PredType.MERGE)
# ASR (immediate, unpredicated)
sveBinImmInst('asr', 'AsrImmUnpred', 'SimdAluOp', unsignedTypes, asrCode)
# ASR (vectors)
sveBinInst('asr', 'AsrPred', 'SimdAluOp', unsignedTypes, asrCode,
PredType.MERGE, True)
# ASR (wide elements, predicated)
sveShiftByWideElemsInst('asr', 'AsrWidePred', 'SimdAluOp', unsignedTypes,
asrCode, PredType.MERGE)
# ASR (wide elements, unpredicated)
sveShiftByWideElemsInst('asr', 'AsrWideUnpred', 'SimdAluOp', unsignedTypes,
asrCode)
# ASRD
asrdCode = '''
Element element1 = srcElem1;
Element shift = srcElem2;
if (srcElem1 < 0) {
Element tmp = ((1L << shift) - 1L);
if (tmp == -1L) {
element1 = 0;
} else {
element1 = element1 + tmp;
}
}
destElem = (element1 >> shift);
'''
sveBinImmInst('asrd', 'Asrd', 'SimdAluOp', signedTypes, asrdCode,
PredType.MERGE)
# ASRR
asrrCode = '''
int sign_bit = bits(srcElem2, sizeof(Element) * 8 - 1);
if (srcElem1 == 0) {
destElem = srcElem2;
} else if (srcElem1 >= sizeof(Element) * 8) {
destElem = sign_bit ? std::numeric_limits<Element>::max() : 0;
} else {
destElem = srcElem2 >> srcElem1;
if (sign_bit) {
destElem |= ~mask(sizeof(Element) * 8 - srcElem1);
}
}
'''
sveBinInst('asrr', 'Asrr', 'SimdAluOp', unsignedTypes, asrrCode,
PredType.MERGE, True)
# BIC (vectors, predicated)
bicCode = 'destElem = srcElem1 & ~srcElem2;'
sveBinInst('bic', 'BicPred', 'SimdAluOp', unsignedTypes, bicCode,
PredType.MERGE, True)
# BIC (vectors, unpredicated)
sveBinInst('bic', 'BicUnpred', 'SimdAluOp', unsignedTypes, bicCode)
# BIC, BICS (predicates)
bicCode = 'destElem = srcElem1 && !srcElem2;'
svePredLogicalInst('bic', 'PredBic', 'SimdPredAluOp', ('uint8_t',),
bicCode)
svePredLogicalInst('bics', 'PredBics', 'SimdPredAluOp', ('uint8_t',),
bicCode, isFlagSetting=True)
# BRKA (merging)
svePartBrkInst('brka', 'Brkam', 'SimdPredAluOp', isFlagSetting = False,
predType = PredType.MERGE, whenBrk = Break.After)
# BRKA (zeroing)
svePartBrkInst('brka', 'Brkaz', 'SimdPredAluOp', isFlagSetting = False,
predType = PredType.ZERO, whenBrk = Break.After)
# BRKAS
svePartBrkInst('brkas', 'Brkas', 'SimdPredAluOp', isFlagSetting = True,
predType = PredType.ZERO, whenBrk = Break.After)
# BRKB (merging)
svePartBrkInst('brkb', 'Brkbm', 'SimdPredAluOp', isFlagSetting = False,
predType = PredType.MERGE, whenBrk = Break.Before)
# BRKB (zeroging)
svePartBrkInst('brkb', 'Brkbz', 'SimdPredAluOp', isFlagSetting = False,
predType = PredType.ZERO, whenBrk = Break.Before)
# BRKBS
svePartBrkInst('brkbs', 'Brkbs', 'SimdPredAluOp', isFlagSetting = True,
predType = PredType.ZERO, whenBrk = Break.Before)
# BRKN
svePartBrkPropNextInst('brkn', 'Brkn', 'SimdPredAluOp',
isFlagSetting = False)
# BRKNS
svePartBrkPropNextInst('brkns', 'Brkns', 'SimdPredAluOp',
isFlagSetting = True)
# BRKPA
svePartBrkPropPrevInst('brkpa', 'Brkpa', 'SimdPredAluOp',
isFlagSetting = False, whenBrk = Break.After)
# BRKPAS
svePartBrkPropPrevInst('brkpas', 'Brkpas', 'SimdPredAluOp',
isFlagSetting = True, whenBrk = Break.After)
# BRKPB
svePartBrkPropPrevInst('brkpb', 'Brkpb', 'SimdPredAluOp',
isFlagSetting = False, whenBrk = Break.Before)
# BRKPBS
svePartBrkPropPrevInst('brkpbs', 'Brkpbs', 'SimdPredAluOp',
isFlagSetting = True, whenBrk = Break.Before)
# CLASTA (scalar)
clastaCode = '''
last++;
if (last >= eCount)
last = 0;
destElem = AA64FpOp1_x[last];'''
sveSelectInst('clasta', 'Clasta', 'SimdAluOp', unsignedTypes, clastaCode,
isCond = True, destType = DstRegType.Scalar)
# CLASTA (SIMD&FP scalar)
sveSelectInst('clasta', 'Clastaf', 'SimdAluOp', unsignedTypes, clastaCode,
isCond = True, destType = DstRegType.SimdFpScalar)
# CLASTA (vector)
sveSelectInst('clasta', 'Clastav', 'SimdAluOp', unsignedTypes, clastaCode,
isCond = True, destType = DstRegType.Vector)
# CLASTB (scalar)
clastbCode = '''
destElem = AA64FpOp1_x[last];'''
sveSelectInst('clastb', 'Clastb', 'SimdAluOp', unsignedTypes, clastbCode,
isCond = True, destType = DstRegType.Scalar)
# CLASTB (SIMD&FP scalar)
sveSelectInst('clastb', 'Clastbf', 'SimdAluOp', unsignedTypes, clastbCode,
isCond = True, destType = DstRegType.SimdFpScalar)
# CLASTB (vectors)
sveSelectInst('clastb', 'Clastbv', 'SimdAluOp', unsignedTypes, clastbCode,
isCond = True, destType = DstRegType.Vector)
# CLS
clsCode = '''
destElem = 0;
Element val = srcElem1;
if (val < 0) {
val <<= 1;
while (val < 0) {
destElem++;
val <<= 1;
}
} else {
val <<= 1;
while (val >= 0 && destElem < sizeof(Element) * 8 - 1) {
destElem++;
val <<= 1;
}
}
'''
sveUnaryInst('cls', 'Cls', 'SimdAluOp', signedTypes, clsCode,
PredType.MERGE)
# CLZ
clzCode = '''
destElem = 0;
Element val = srcElem1;
while (val >= 0 && destElem < sizeof(Element) * 8) {
destElem++;
val <<= 1;
}
'''
sveUnaryInst('clz', 'Clz', 'SimdAluOp', signedTypes, clzCode,
PredType.MERGE)
# CMPEQ (immediate)
cmpeqCode = '''
destElem = (srcElem1 == srcElem2);
'''
sveIntCmpImmInst('cmpeq', 'Cmpeqi', 'SimdCmpOp', unsignedTypes, cmpeqCode)
# CMPEQ (vectors)
sveIntCmpInst('cmpeq', 'Cmpeq', 'SimdCmpOp', unsignedTypes, cmpeqCode)
# CMPEQ (wide elements)
sveIntCmpInst('cmpeq', 'Cmpeqw', 'SimdCmpOp', smallUnsignedTypes,
cmpeqCode, True)
# CMPGE (immediate)
cmpgeCode = '''
destElem = (srcElem1 >= srcElem2);
'''
sveIntCmpImmInst('cmpge', 'Cmpgei', 'SimdCmpOp', signedTypes, cmpgeCode)
# CMPGE (vectors)
sveIntCmpInst('cmpge', 'Cmpge', 'SimdCmpOp', signedTypes, cmpgeCode)
# CMPGE (wide elements)
sveIntCmpInst('cmpge', 'Cmpgew', 'SimdCmpOp', smallSignedTypes,
cmpgeCode, True)
# CMPGT (immediate)
cmpgtCode = '''
destElem = (srcElem1 > srcElem2);
'''
sveIntCmpImmInst('cmpge', 'Cmpgti', 'SimdCmpOp', signedTypes, cmpgtCode)
# CMPGT (vectors)
sveIntCmpInst('cmpge', 'Cmpgt', 'SimdCmpOp', signedTypes, cmpgtCode)
# CMPGT (wide elements)
sveIntCmpInst('cmpge', 'Cmpgtw', 'SimdCmpOp', smallSignedTypes,
cmpgtCode, True)
# CMPHI (immediate)
sveIntCmpImmInst('cmphi', 'Cmphii', 'SimdCmpOp', unsignedTypes, cmpgtCode)
# CMPHI (vectors)
sveIntCmpInst('cmphi', 'Cmphi', 'SimdCmpOp', unsignedTypes, cmpgtCode)
# CMPHI (wide elements)
sveIntCmpInst('cmphi', 'Cmphiw', 'SimdCmpOp', smallUnsignedTypes,
cmpgtCode, True)
# CMPHS (immediate)
sveIntCmpImmInst('cmphs', 'Cmphsi', 'SimdCmpOp', unsignedTypes, cmpgeCode)
# CMPHS (vectors)
sveIntCmpInst('cmphs', 'Cmphs', 'SimdCmpOp', unsignedTypes, cmpgeCode)
# CMPHS (wide elements)
sveIntCmpInst('cmphs', 'Cmphsw', 'SimdCmpOp', smallUnsignedTypes,
cmpgeCode, True)
# CMPLE (immediate)
cmpleCode = '''
destElem = (srcElem1 <= srcElem2);
'''
sveIntCmpImmInst('cmple', 'Cmplei', 'SimdCmpOp', signedTypes, cmpleCode)
# CMPLE (wide elements)
sveIntCmpInst('cmple', 'Cmplew', 'SimdCmpOp', smallSignedTypes,
cmpleCode, True)
# CMPLO (immediate)
cmpltCode = '''
destElem = (srcElem1 < srcElem2);
'''
sveIntCmpImmInst('cmplo', 'Cmploi', 'SimdCmpOp', unsignedTypes, cmpltCode)
# CMPLO (wide elements)
sveIntCmpInst('cmplo', 'Cmplow', 'SimdCmpOp', smallUnsignedTypes,
cmpltCode, True)
# CMPLS (immediate)
sveIntCmpImmInst('cmpls', 'Cmplsi', 'SimdCmpOp', unsignedTypes, cmpleCode)
# CMPLS (wide elements)
sveIntCmpInst('cmpls', 'Cmplsw', 'SimdCmpOp', smallUnsignedTypes,
cmpleCode, True)
# CMPLT (immediate)
sveIntCmpImmInst('cmplt', 'Cmplti', 'SimdCmpOp', signedTypes, cmpltCode)
# CMPLT (wide elements)
sveIntCmpInst('cmplt', 'Cmpltw', 'SimdCmpOp', smallSignedTypes,
cmpltCode, True)
# CMPNE (immediate)
cmpneCode = '''
destElem = (srcElem1 != srcElem2);
'''
sveIntCmpImmInst('cmpeq', 'Cmpnei', 'SimdCmpOp', unsignedTypes, cmpneCode)
# CMPNE (vectors)
sveIntCmpInst('cmpeq', 'Cmpne', 'SimdCmpOp', unsignedTypes, cmpneCode)
# CMPNE (wide elements)
sveIntCmpInst('cmpeq', 'Cmpnew', 'SimdCmpOp', smallUnsignedTypes,
cmpneCode, True)
# CNOT
cnotCode = '''
destElem = srcElem1?0:1;
'''
sveUnaryInst('cnot', 'Cnot', 'SimdAluOp', unsignedTypes, cnotCode,
PredType.MERGE)
# CNT
cntCode = '''
destElem = 0;
Element val = srcElem1;
while (val) {
destElem += val & 0x1;
val >>= 1;
}
'''
sveUnaryInst('cnt', 'Cnt', 'SimdAluOp', unsignedTypes, cntCode,
PredType.MERGE)
# CNTB, CNTD, CNTH, CNTW
cntxCode = '''
destElem = (count * imm);
'''
sveElemCountInst('cnt', 'Cntx', 'SimdAluOp', unsignedTypes, cntxCode,
destType = DestType.Scalar, dstIs32b = False, dstAcc = False)
# COMPACT
sveCompactInst('compact', 'Compact', 'SimdPredAluOp',
('uint32_t', 'uint64_t'))
# CPY (immediate)
dupCode = 'destElem = srcElem1;'
sveWideImmInst('cpy', 'CpyImmMerge', 'SimdAluOp', unsignedTypes, dupCode,
predType=PredType.MERGE, isUnary=True)
sveWideImmInst('cpy', 'CpyImmZero', 'SimdAluOp', unsignedTypes, dupCode,
predType=PredType.ZERO, isUnary=True)
# CPY (scalar)
sveUnaryInst('cpy', 'CpyScalar', 'SimdAluOp', unsignedTypes, dupCode,
PredType.MERGE, srcRegType=SrcRegType.Scalar)
# CPY (SIMD&FP scalar)
sveUnaryInst('cpy', 'CpySimdFpScalar', 'SimdAluOp', unsignedTypes, dupCode,
PredType.MERGE, srcRegType=SrcRegType.SimdFpScalar)
# CNTP
svePredCountPredInst('cntp', 'Cntp', 'SimdAluOp', unsignedTypes)
# CTERMEQ
cteqCode = '''
destElem = srcElem1 == srcElem2;
'''
sveCompTermInst('ctermeq', 'Ctermeq', 'IntAluOp',
['uint32_t', 'uint64_t'], cteqCode)
# CTERMNE
ctneCode = '''
destElem = srcElem1 != srcElem2;
'''
sveCompTermInst('ctermne', 'Ctermne', 'IntAluOp',
['uint32_t', 'uint64_t'], ctneCode)
# DECB, DECH, DECW, DECD (scalar)
decxCode = '''
destElem = srcElem1 - (count * imm);
'''
sveElemCountInst('dec', 'Dec', 'SimdAluOp', unsignedTypes, decxCode,
destType = DestType.Scalar, dstIs32b = False)
# DECH, DECW, DECD (vector)
sveElemCountInst('dec', 'Decv', 'SimdAluOp', bigUnsignedTypes, decxCode,
destType = DestType.Vector, dstIs32b = False)
# DECP (scalar)
decpCode = '''
XDest = XDest - count;
'''
svePredCountInst('decp', 'Decp', 'SimdAluOp', unsignedTypes, decpCode,
DestType.Scalar, SrcSize.Src64bit)
# DECP (vector)
decpvCode = '''
destElem = srcElem - count;
'''
svePredCountInst('decp', 'Decpv', 'SimdAluOp', unsignedTypes, decpvCode,
DestType.Vector)
# DUP (immediate)
sveWideImmInst('dup', 'DupImm', 'SimdAluOp', unsignedTypes, dupCode,
isUnary=True)
# DUP (indexed)
sveDupIndexInst('mov', 'DupIdx', 'SimdAluOp',
list(unsignedTypes) + ['__uint128_t'])
# DUP (scalar)
sveUnaryInst('dup', 'DupScalar', 'SimdAluOp', unsignedTypes, dupCode,
PredType.NONE, srcRegType=SrcRegType.Scalar)
# DUPM
sveWideImmInst('dupm', 'Dupm', 'SimdAluOp', unsignedTypes, dupCode,
isUnary=True)
# EOR (immediate)
eorCode = 'destElem = srcElem1 ^ srcElem2;'
sveWideImmInst('eor', 'EorImm', 'SimdAluOp', ('uint64_t',), eorCode)
# EOR (vectors, predicated)
sveBinInst('eor', 'EorPred', 'SimdAluOp', unsignedTypes, eorCode,
PredType.MERGE, True)
# EOR (vectors, unpredicated)
eorCode = 'destElem = srcElem1 ^ srcElem2;'
sveBinInst('eor', 'EorUnpred', 'SimdAluOp', ('uint64_t',), eorCode)
# EOR, EORS (predicates)
svePredLogicalInst('eor', 'PredEor', 'SimdPredAluOp', ('uint8_t',),
eorCode)
svePredLogicalInst('eors', 'PredEors', 'SimdPredAluOp', ('uint8_t',),
eorCode, isFlagSetting=True)
# EORV
eorvCode = 'destElem ^= srcElem1;'
sveAssocReducInst('eorv', 'Eorv', 'SimdReduceAluOp', unsignedTypes,
eorvCode, '0')
# EXT
sveExtInst('ext', 'Ext', 'SimdAluOp')
# FABD
fpOp = '''
FPSCR fpscr = FpscrExc & ~FpscrAhpMask;
destElem = %s;
FpscrExc = fpscr;
'''
fabdCode = fpOp % 'fplibAbs<Element>(fplibSub(srcElem1, srcElem2, fpscr))'
sveBinInst('fabd', 'Fabd', 'SimdFloatAddOp', floatTypes, fabdCode,
PredType.MERGE, True)
# FABS
fabsCode = 'destElem = fplibAbs<Element>(srcElem1);'
sveUnaryInst('fabs', 'Fabs', 'SimdFloatAluOp', fpTypes, fabsCode,
PredType.MERGE)
# FACGE
fpCmpAbsOp = fpOp % ('fplibCompare%s<Element>(fplibAbs<Element>(srcElem1),'
' fplibAbs<Element>(srcElem2), fpscr)')
facgeCode = fpCmpAbsOp % 'GE'
sveCmpInst('facge', 'Facge', 'SimdFloatCmpOp', fpTypes, facgeCode)
# FACGT
facgtCode = fpCmpAbsOp % 'GT'
sveCmpInst('facgt', 'Facgt', 'SimdFloatCmpOp', fpTypes, facgtCode)
# FADD (immediate)
fpBinOp = fpOp % 'fplib%s<Element>(srcElem1, srcElem2, fpscr)'
faddCode = fpBinOp % 'Add'
sveBinImmInst('fadd', 'FaddImm', 'SimdFloatAddOp', floatTypes, faddCode,
PredType.MERGE)
# FADD (vectors, predicated)
sveBinInst('fadd', 'FaddPred', 'SimdFloatAddOp', floatTypes, faddCode,
PredType.MERGE, True)
# FADD (vectors, unpredicated)
sveBinInst('fadd', 'FaddUnpred', 'SimdFloatAddOp', floatTypes, faddCode)
# FADDA
fpAddaOp = '''
FPSCR fpscr = FpscrExc & ~FpscrAhpMask;
destElem = fplibAdd<Element>(destElem, srcElem1, fpscr);
FpscrExc = FpscrExc | fpscr;
'''
sveOrderedReduction('fadda', 'Fadda', 'SimdFloatReduceAddOp', floatTypes,
fpAddaOp)
# FADDV
fpReduceOp = '''
FPSCR fpscr = FpscrExc & ~FpscrAhpMask;
destElem = fplib%s<Element>(srcElem1, srcElem2, fpscr);
FpscrExc = FpscrExc | fpscr;
'''
faddvCode = fpReduceOp % 'Add'
sveNonAssocReducInst('faddv', 'Faddv', 'SimdFloatReduceAddOp', floatTypes,
faddvCode, '0')
# FCADD
sveComplexAddInst('fcadd','Fcadd', 'SimdFloatAddOp', fpTypes)
# FCMEQ (vectors)
fpCmpOp = fpOp % ('fplibCompare%s<Element>(srcElem1, srcElem2, fpscr)')
fcmeqCode = fpCmpOp % 'EQ'
sveCmpInst('fcmeq', 'Fcmeq', 'SimdFloatCmpOp', fpTypes, fcmeqCode)
# FCMEQ (zero)
fpCmpZeroOp = fpOp % 'fplibCompare%s<Element>(srcElem1, 0, fpscr)'
fcmeqZeroCode = fpCmpZeroOp % 'EQ'
sveCmpInst('fcmeq', 'FcmeqZero', 'SimdFloatCmpOp', fpTypes, fcmeqZeroCode,
True)
# FCMGE (vectors)
fcmgeCode = fpCmpOp % 'GE'
sveCmpInst('fcmge', 'Fcmge', 'SimdFloatCmpOp', fpTypes, fcmgeCode)
# FCMGE (zero)
fcmgeZeroCode = fpCmpZeroOp % 'GE'
sveCmpInst('fcmge', 'FcmgeZero', 'SimdFloatCmpOp', fpTypes, fcmgeZeroCode,
True)
# FCMGT (vectors)
fcmgtCode = fpCmpOp % 'GT'
sveCmpInst('fcmgt', 'Fcmgt', 'SimdFloatCmpOp', fpTypes, fcmgtCode)
# FCMGT (zero)
fcmgtZeroCode = fpCmpZeroOp % 'GT'
sveCmpInst('fcmgt', 'FcmgtZero', 'SimdFloatCmpOp', fpTypes, fcmgtZeroCode,
True)
# FCMLE (zero)
fpCmpRevZeroOp = fpOp % ('fplibCompare%s<Element>(0, srcElem1, fpscr)')
fcmleZeroCode = fpCmpRevZeroOp % 'GE'
sveCmpInst('fcmle', 'FcmleZero', 'SimdFloatCmpOp', fpTypes, fcmleZeroCode,
True)
# FCMLT (zero)
fcmltZeroCode = fpCmpRevZeroOp % 'GT'
sveCmpInst('fcmlt', 'FcmltZero', 'SimdFloatCmpOp', fpTypes, fcmltZeroCode,
True)
# FCMNE (vectors)
fcmneCode = fpOp % ('!fplibCompareEQ<Element>(srcElem1, srcElem2, fpscr)')
sveCmpInst('fcmne', 'Fcmne', 'SimdFloatCmpOp', fpTypes, fcmneCode)
# FCMNE (zero)
fcmneZeroCode = fpOp % ('!fplibCompareEQ<Element>(srcElem1, 0, fpscr)')
sveCmpInst('fcmne', 'FcmneZero', 'SimdFloatCmpOp', fpTypes, fcmneZeroCode,
True)
# FCMUO (vectors)
fcmuoCode = fpCmpOp % 'UN'
sveCmpInst('fcmuo', 'Fcmuo', 'SimdFloatCmpOp', fpTypes, fcmuoCode)
# FCMLA (indexed)
sveComplexMulAddInst('fcmla', 'Fcmlai', 'SimdFloatMultAccOp',
fpTypes[:2], predType = PredType.NONE)
# FCMLA (vectors)
sveComplexMulAddInst('fcmla', 'Fcmlav', 'SimdFloatMultAccOp',
fpTypes, predType = PredType.MERGE)
# FCPY
sveWideImmInst('fcpy', 'Fcpy', 'SimdAluOp', unsignedTypes, dupCode,
predType=PredType.MERGE, isUnary=True)
# FCVT
fcvtCode = fpOp % ('fplibConvert<SElement, DElement>('
'srcElem1, FPCRRounding(fpscr), fpscr)')
sveCvtInst('fcvt', 'FcvtNarrow', 'SimdCvtOp',
('uint32_t, uint16_t',
'uint64_t, uint16_t',
'uint64_t, uint32_t'),
fcvtCode, CvtDir.Narrow)
sveCvtInst('fcvt', 'FcvtWiden', 'SimdCvtOp',
('uint16_t, uint32_t',
'uint16_t, uint64_t',
'uint32_t, uint64_t'),
fcvtCode, CvtDir.Widen)
# FCVTZS
fcvtIntCode = fpOp % ('fplibFPToFixed<SElement, DElement>('
'srcElem1, %s, %s, %s, fpscr)')
fcvtzsCode = fcvtIntCode % ('0', 'false', 'FPRounding_ZERO')
sveCvtInst('fcvtzs', 'FcvtzsNarrow', 'SimdCvtOp',
('uint16_t, uint16_t',
'uint32_t, uint32_t',
'uint64_t, uint32_t',
'uint64_t, uint64_t'),
fcvtzsCode, CvtDir.Narrow, signed=True)
sveCvtInst('fcvtzs', 'FcvtzsWiden', 'SimdCvtOp',
('uint16_t, uint32_t',
'uint16_t, uint64_t',
'uint32_t, uint64_t'),
fcvtzsCode, CvtDir.Widen)
# FCVTZU
fcvtzuCode = fcvtIntCode % ('0', 'true', 'FPRounding_ZERO')
sveCvtInst('fcvtzu', 'FcvtzuNarrow', 'SimdCvtOp',
('uint16_t, uint16_t',
'uint32_t, uint32_t',
'uint64_t, uint32_t',
'uint64_t, uint64_t'),
fcvtzuCode, CvtDir.Narrow)
sveCvtInst('fcvtzu', 'FcvtzuWiden', 'SimdCvtOp',
('uint16_t, uint32_t',
'uint16_t, uint64_t',
'uint32_t, uint64_t'),
fcvtzuCode, CvtDir.Widen)
# FDIV
fdivCode = fpBinOp % 'Div'
sveBinInst('fdiv', 'Fdiv', 'SimdFloatDivOp', floatTypes, fdivCode,
PredType.MERGE, True)
# FDIVR
fpBinRevOp = fpOp % 'fplib%s<Element>(srcElem2, srcElem1, fpscr)'
fdivrCode = fpBinRevOp % 'Div'
sveBinInst('fdivr', 'Fdivr', 'SimdFloatDivOp', floatTypes, fdivrCode,
PredType.MERGE, True)
# FDUP
sveWideImmInst('fdup', 'Fdup', 'SimdFloatAluOp', floatTypes, dupCode,
isUnary=True)
# FEXPA
fexpaCode = 'destElem = fplibExpA<Element>(srcElem1);'
sveUnaryInst('fexpa', 'Fexpa', 'SimdFloatAluOp', fpTypes, fexpaCode)
# FMAD
fmadCode = fpOp % ('fplibMulAdd<Element>('
'srcElem1, destElem, srcElem2, fpscr)')
sveTerInst('fmad', 'Fmad', 'SimdFloatMultAccOp', floatTypes, fmadCode,
PredType.MERGE)
# FMAX (immediate)
fmaxCode = fpBinOp % 'Max'
sveBinImmInst('fmax', 'FmaxImm', 'SimdFloatCmpOp', floatTypes, fmaxCode,
PredType.MERGE)
# FMAX (vectors)
sveBinInst('fmax', 'Fmax', 'SimdFloatCmpOp', floatTypes, fmaxCode,
PredType.MERGE, True)
# FMAXNM (immediate)
fmaxnmCode = fpBinOp % 'MaxNum'
sveBinImmInst('fmaxnm', 'FmaxnmImm', 'SimdFloatCmpOp', floatTypes,
fmaxnmCode, PredType.MERGE)
# FMAXNM (vectors)
sveBinInst('fmaxnm', 'Fmaxnm', 'SimdFloatCmpOp', floatTypes, fmaxnmCode,
PredType.MERGE, True)
# FMAXNMV
fmaxnmvCode = fpReduceOp % 'MaxNum'
sveNonAssocReducInst('fmaxnmv', 'Fmaxnmv', 'SimdFloatReduceCmpOp',
floatTypes, fmaxnmvCode, 'fplibDefaultNaN<Element>()')
# FMAXV
fmaxvCode = fpReduceOp % 'Max'
sveNonAssocReducInst('fmaxv', 'Fmaxv', 'SimdFloatReduceCmpOp', floatTypes,
fmaxvCode, 'fplibInfinity<Element>(1)')
# FMIN (immediate)
fminCode = fpBinOp % 'Min'
sveBinImmInst('fmin', 'FminImm', 'SimdFloatCmpOp', floatTypes, fminCode,
PredType.MERGE)
# FMIN (vectors)
sveBinInst('fmin', 'Fmin', 'SimdFloatCmpOp', floatTypes, fminCode,
PredType.MERGE, True)
# FMINNM (immediate)
fminnmCode = fpBinOp % 'MinNum'
sveBinImmInst('fminnm', 'FminnmImm', 'SimdFloatCmpOp', floatTypes,
fminnmCode, PredType.MERGE)
# FMINNM (vectors)
sveBinInst('fminnm', 'Fminnm', 'SimdFloatCmpOp', floatTypes, fminnmCode,
PredType.MERGE, True)
# FMINNMV
fminnmvCode = fpReduceOp % 'MinNum'
sveNonAssocReducInst('fminnmv', 'Fminnmv', 'SimdFloatReduceCmpOp',
floatTypes, fminnmvCode, 'fplibDefaultNaN<Element>()')
# FMINV
fminvCode = fpReduceOp % 'Min'
sveNonAssocReducInst('fminv', 'Fminv', 'SimdFloatReduceCmpOp', floatTypes,
fminvCode, 'fplibInfinity<Element>(0)')
fmlaCode = fpOp % ('fplibMulAdd<Element>('
'destElem, srcElem1, srcElem2, fpscr)')
# FMLA (indexed)
sveTerIdxInst('fmla', 'FmlaIdx', 'SimdFloatMultAccOp', floatTypes,
fmlaCode, PredType.MERGE)
# FMLA (vectors)
sveTerInst('fmla', 'Fmla', 'SimdFloatMultAccOp', floatTypes, fmlaCode,
PredType.MERGE)
fmlsCode = fpOp % ('fplibMulAdd<Element>(destElem, '
'fplibNeg<Element>(srcElem1), srcElem2, fpscr)')
# FMLS (indexed)
sveTerIdxInst('fmls', 'FmlsIdx', 'SimdFloatMultAccOp', floatTypes,
fmlsCode, PredType.MERGE)
# FMLS (vectors)
sveTerInst('fmls', 'Fmls', 'SimdFloatMultAccOp', floatTypes, fmlsCode,
PredType.MERGE)
# FMSB
fmsbCode = fpOp % ('fplibMulAdd<Element>(srcElem1, '
'fplibNeg<Element>(destElem), srcElem2, fpscr)')
sveTerInst('fmsb', 'Fmsb', 'SimdFloatMultAccOp', floatTypes, fmsbCode,
PredType.MERGE)
# FMUL (immediate)
fpBinOp = fpOp % 'fplib%s<Element>(srcElem1, srcElem2, fpscr)'
fmulCode = fpBinOp % 'Mul'
sveBinImmInst('fmul', 'FmulImm', 'SimdFloatMultOp', floatTypes, fmulCode,
PredType.MERGE)
# TODO: FMUL (indexed)
# FMUL (vectors, predicated)
fmulCode = fpBinOp % 'Mul'
sveBinInst('fmul', 'FmulPred', 'SimdFloatMultOp', floatTypes, fmulCode,
PredType.MERGE, True)
# FMUL (vectors, unpredicated)
sveBinInst('fmul', 'FmulUnpred', 'SimdFloatMultOp', floatTypes, fmulCode)
# FMUL (indexed)
sveBinIdxInst('fmul', 'FmulIdx', 'SimdFloatMultOp', floatTypes, fmulCode)
# FMULX
fmulxCode = fpBinOp % 'MulX'
sveBinInst('fmulx', 'Fmulx', 'SimdFloatMultOp', floatTypes, fmulxCode,
PredType.MERGE, True)
# FNEG
fnegCode = 'destElem = fplibNeg<Element>(srcElem1);'
sveUnaryInst('fneg', 'Fneg', 'SimdFloatAluOp', fpTypes, fnegCode,
PredType.MERGE)
# FNMAD
fnmadCode = fpOp % ('fplibMulAdd<Element>('
'fplibNeg<Element>(srcElem1), '
'fplibNeg<Element>(destElem), srcElem2, fpscr)')
sveTerInst('fnmad', 'Fnmad', 'SimdFloatMultAccOp', floatTypes, fnmadCode,
PredType.MERGE)
# FNMLA
fnmlaCode = fpOp % ('fplibMulAdd<Element>('
'fplibNeg<Element>(destElem), '
'fplibNeg<Element>(srcElem1), srcElem2, fpscr)')
sveTerInst('fnmla', 'Fnmla', 'SimdFloatMultAccOp', floatTypes, fnmlaCode,
PredType.MERGE)
# FNMLS
fnmlsCode = fpOp % ('fplibMulAdd<Element>('
'fplibNeg<Element>(destElem), srcElem1, srcElem2, '
'fpscr)')
sveTerInst('fnmls', 'Fnmls', 'SimdFloatMultAccOp', floatTypes, fnmlsCode,
PredType.MERGE)
# FNMSB
fnmsbCode = fpOp % ('fplibMulAdd<Element>('
'fplibNeg<Element>(srcElem1), destElem, srcElem2, '
'fpscr)')
sveTerInst('fnmsb', 'Fnmsb', 'SimdFloatMultAccOp', floatTypes, fnmsbCode,
PredType.MERGE)
# FRECPE
frecpeCode = fpOp % 'fplibRecipEstimate<Element>(srcElem1, fpscr)'
sveUnaryInst('frecpe', 'Frecpe', 'SimdFloatMultAccOp', floatTypes,
frecpeCode)
# FRECPS
frecpsCode = fpBinOp % 'RecipStepFused'
sveBinInst('frecps', 'Frecps', 'SimdFloatMultAccOp', floatTypes,
frecpsCode)
# FRECPX
frecpxCode = fpOp % "fplibRecpX<Element>(srcElem1, fpscr)"
sveUnaryInst('frecpx', 'Frecpx', 'SimdFloatMultAccOp', floatTypes,
frecpxCode, PredType.MERGE)
# FRINTA
frintCode = fpOp % 'fplibRoundInt<Element>(srcElem1, %s, %s, fpscr)'
frintaCode = frintCode % ('FPRounding_TIEAWAY', 'false')
sveUnaryInst('frinta', 'Frinta', 'SimdCvtOp', floatTypes, frintaCode,
PredType.MERGE)
# FRINTI
frintiCode = frintCode % ('FPCRRounding(fpscr)', 'false')
sveUnaryInst('frinti', 'Frinti', 'SimdCvtOp', floatTypes, frintiCode,
PredType.MERGE)
# FRINTM
frintmCode = frintCode % ('FPRounding_NEGINF', 'false')
sveUnaryInst('frintm', 'Frintm', 'SimdCvtOp', floatTypes, frintmCode,
PredType.MERGE)
# FRINTN
frintnCode = frintCode % ('FPRounding_TIEEVEN', 'false')
sveUnaryInst('frintn', 'Frintn', 'SimdCvtOp', floatTypes, frintnCode,
PredType.MERGE)
# FRINTP
frintpCode = frintCode % ('FPRounding_POSINF', 'false')
sveUnaryInst('frintp', 'Frintp', 'SimdCvtOp', floatTypes, frintpCode,
PredType.MERGE)
# FRINTX
frintxCode = frintCode % ('FPCRRounding(fpscr)', 'true')
sveUnaryInst('frintx', 'Frintx', 'SimdCvtOp', floatTypes, frintxCode,
PredType.MERGE)
# FRINTZ
frintzCode = frintCode % ('FPRounding_ZERO', 'false')
sveUnaryInst('frintz', 'Frintz', 'SimdCvtOp', floatTypes, frintzCode,
PredType.MERGE)
# FRSQRTE
frsqrteCode = fpOp % 'fplibRSqrtEstimate<Element>(srcElem1, fpscr)'
sveUnaryInst('frsqrte', 'Frsqrte', 'SimdFloatSqrtOp', floatTypes,
frsqrteCode)
# FRSQRTS
frsqrtsCode = fpBinOp % 'RSqrtStepFused'
sveBinInst('frsqrts', 'Frsqrts', 'SimdFloatMiscOp', floatTypes,
frsqrtsCode)
# FSCALE
fscaleCode = fpBinOp % 'Scale'
sveBinInst('fscale', 'Fscale', 'SimdFloatMiscOp', floatTypes, fscaleCode,
PredType.MERGE, True)
# FSQRT
fsqrtCode = fpOp % "fplibSqrt<Element>(srcElem1, fpscr)"
sveUnaryInst('fsqrt', 'Fsqrt', 'SimdFloatSqrtOp', floatTypes, fsqrtCode,
PredType.MERGE)
# FSUB (immediate)
fsubCode = fpBinOp % 'Sub'
sveBinImmInst('fsub', 'FsubImm', 'SimdFloatAddOp', floatTypes, fsubCode,
PredType.MERGE)
# FSUB (vectors, predicated)
sveBinInst('fsub', 'FsubPred', 'SimdFloatAddOp', floatTypes, fsubCode,
PredType.MERGE, True)
# FSUB (vectors, unpredicated)
sveBinInst('fsub', 'FsubUnpred', 'SimdFloatAddOp', floatTypes, fsubCode)
# FSUBR (immediate)
fsubrCode = fpBinRevOp % 'Sub'
sveBinImmInst('fsubr', 'FsubrImm', 'SimdFloatAddOp', floatTypes, fsubrCode,
PredType.MERGE)
# FSUBR (vectors)
sveBinInst('fsubr', 'Fsubr', 'SimdFloatAddOp', floatTypes, fsubrCode,
PredType.MERGE, True)
# FTMAD
ftmadCode = fpOp % ('fplibTrigMulAdd<Element>('
'srcElem3, destElem, srcElem2, fpscr)')
sveTerImmInst('ftmad', 'Ftmad', 'SimdFloatMultAccOp', floatTypes,
ftmadCode)
# FTSMUL
ftsmulCode = fpBinOp % 'TrigSMul'
sveBinInst('ftsmul', 'Ftsmul', 'SimdFloatMiscOp', floatTypes, ftsmulCode)
# FTSSEL
ftsselCode = fpBinOp % 'TrigSSel'
sveBinInst('ftssel', 'Ftssel', 'SimdFloatMultOp', floatTypes, ftsselCode)
# INCB, INCH, INCW, INCD (scalar)
incxCode = '''
destElem = srcElem1 + (count * imm);
'''
sveElemCountInst('inc', 'Inc', 'SimdAluOp', unsignedTypes, incxCode,
destType = DestType.Scalar, dstIs32b = False)
# INCH, INCW, INCD (vector)
sveElemCountInst('inc', 'Incv', 'SimdAluOp', bigUnsignedTypes, incxCode,
destType = DestType.Vector, dstIs32b = False)
# INCP (scalar)
incpCode = '''
XDest = XDest + count;
'''
svePredCountInst('incp', 'Incp', 'SimdAluOp', unsignedTypes, incpCode,
DestType.Scalar, SrcSize.Src64bit)
# INCP (vector)
incpvCode = '''
destElem = srcElem + count;
'''
svePredCountInst('incp', 'Incpv', 'SimdAluOp', unsignedTypes, incpvCode,
DestType.Vector)
# INDEX (immediate, scalar)
sveIndex(IndexFormat.ImmReg)
# INDEX (immediates)
sveIndex(IndexFormat.ImmImm)
# INDEX (scalar, immediate)
sveIndex(IndexFormat.RegImm)
# INDEX (scalars)
sveIndex(IndexFormat.RegReg)
# INSR (scalar)
sveShiftAndInsertInst('insr', 'Insr', 'SimdAluOp', unsignedTypes,
srcType = SrcRegType.Scalar)
# INSR (SIMD&FP scalar)
sveShiftAndInsertInst('insr', 'Insrf', 'SimdAluOp', unsignedTypes,
srcType = SrcRegType.SimdFpScalar)
# LASTA (scalar)
lastaCode = '''
last++;
if (last >= eCount) {
last = 0;
}
destElem = AA64FpOp1_x[last];'''
sveSelectInst('lasta', 'Lasta', 'SimdAluOp', unsignedTypes, lastaCode,
isCond = False)
# LASTA (SIMD&FP scalar)
sveSelectInst('lasta', 'Lastaf', 'SimdAluOp', unsignedTypes, lastaCode,
isCond = False, destType = DstRegType.SimdFpScalar)
# LASTB (scalar)
lastbCode = '''
if (last < 0) {
last = eCount - 1;
}
destElem = AA64FpOp1_x[last];'''
sveSelectInst('lastb', 'Lastb', 'SimdAluOp', unsignedTypes, lastbCode,
isCond = False)
# LASTB (SIMD&FP scalar)
sveSelectInst('lastb', 'Lastbf', 'SimdAluOp', unsignedTypes, lastbCode,
isCond = False, destType = DstRegType.SimdFpScalar)
# LSL (immediate, predicated)
lslCode = '''
if (srcElem2 == 0) {
destElem = srcElem1;
} else if (srcElem2 >= sizeof(Element) * 8) {
destElem = 0;
} else {
destElem = srcElem1 << srcElem2;
}
'''
sveBinImmInst('lsl', 'LslImmPred', 'SimdAluOp', unsignedTypes, lslCode,
PredType.MERGE)
# LSL (immediate, unpredicated)
sveBinImmInst('lsl', 'LslImmUnpred', 'SimdAluOp', unsignedTypes, lslCode)
# LSL (vectors)
sveBinInst('lsl', 'LslPred', 'SimdAluOp', unsignedTypes, lslCode,
PredType.MERGE, True)
# LSL (wide elements, predicated)
sveShiftByWideElemsInst('lsl', 'LslWidePred', 'SimdAluOp', unsignedTypes,
lslCode, PredType.MERGE)
# LSL (wide elements, unpredicated)
sveShiftByWideElemsInst('lsl', 'LslWideUnpred', 'SimdAluOp', unsignedTypes,
lslCode)
# LSLR
lslrCode = '''
if (srcElem1 == 0) {
destElem = srcElem2;
} else if (srcElem1 >= sizeof(Element) * 8) {
destElem = 0;
} else {
destElem = srcElem2 << srcElem1;
}
'''
sveBinInst('lslr', 'Lslr', 'SimdAluOp', unsignedTypes, lslrCode,
PredType.MERGE, True)
# LSR (immediate, predicated)
lsrCode = '''
if (srcElem2 >= sizeof(Element) * 8) {
destElem = 0;
} else {
destElem = srcElem1 >> srcElem2;
}
'''
sveBinImmInst('lsr', 'LsrImmPred', 'SimdAluOp', unsignedTypes, lsrCode,
PredType.MERGE)
# LSR (immediate, unpredicated)
sveBinImmInst('lsr', 'LsrImmUnpred', 'SimdAluOp', unsignedTypes, lsrCode)
# LSR (vectors)
sveBinInst('lsr', 'LsrPred', 'SimdAluOp', unsignedTypes, lsrCode,
PredType.MERGE, True)
# LSR (wide elements, predicated)
sveShiftByWideElemsInst('lsr', 'LsrWidePred', 'SimdAluOp', unsignedTypes,
lsrCode, PredType.MERGE)
# LSR (wide elements, unpredicated)
sveShiftByWideElemsInst('lsr', 'LsrWideUnpred', 'SimdAluOp', unsignedTypes,
lsrCode)
# LSRR
lsrrCode = '''
if (srcElem1 >= sizeof(Element) * 8) {
destElem = 0;
} else {
destElem = srcElem2 >> srcElem1;
}
'''
sveBinInst('lsrr', 'Lsrr', 'SimdAluOp', unsignedTypes, lsrrCode,
PredType.MERGE, True)
# MAD
madCode = 'destElem = srcElem1 + destElem * srcElem2;'
sveTerInst('mad', 'Mad', 'SimdMultAccOp', signedTypes, madCode)
# MLA
mlaCode = 'destElem += srcElem1 * srcElem2;'
sveTerInst('mla', 'Mla', 'SimdMultAccOp', signedTypes, mlaCode)
# MLS
mlsCode = 'destElem -= srcElem1 * srcElem2;'
sveTerInst('mls', 'Mls', 'SimdMultAccOp', signedTypes, mlsCode)
# MOVPRFX (predicated)
movCode = 'destElem = srcElem1;'
sveUnaryInst('movprfx', 'MovprfxPredM', 'SimdMiscOp', unsignedTypes,
movCode, PredType.MERGE)
sveUnaryInst('movprfx', 'MovprfxPredZ', 'SimdMiscOp', unsignedTypes,
movCode, PredType.ZERO)
# MOVPRFX (unpredicated)
sveUnaryInst('movprfx', 'MovprfxUnpred', 'SimdMiscOp', ('uint64_t',),
movCode)
# MSB
msbCode = 'destElem = srcElem1 - destElem * srcElem2;'
sveTerInst('msb', 'Msb', 'SimdMultAccOp', signedTypes, msbCode)
# MUL (immediate)
mulCode = 'destElem = srcElem1 * srcElem2;'
sveWideImmInst('mul', 'MulImm', 'SimdMultOp', unsignedTypes, mulCode)
# MUL (vectors)
sveBinInst('mul', 'Mul', 'SimdMultOp', unsignedTypes, mulCode,
PredType.MERGE, True)
# NAND, NANDS
nandCode = 'destElem = !(srcElem1 & srcElem2);';
svePredLogicalInst('nand', 'PredNand', 'SimdPredAluOp', ('uint8_t',),
nandCode)
svePredLogicalInst('nands', 'PredNands', 'SimdPredAluOp', ('uint8_t',),
nandCode, isFlagSetting=True)
# NEG
negCode = 'destElem = -srcElem1;'
sveUnaryInst('neg', 'Neg', 'SimdAluOp', signedTypes, negCode,
PredType.MERGE)
# NOR, NORS
norCode = 'destElem = !(srcElem1 | srcElem2);';
svePredLogicalInst('nor', 'PredNor', 'SimdPredAluOp', ('uint8_t',),
norCode)
svePredLogicalInst('nors', 'PredNors', 'SimdPredAluOp', ('uint8_t',),
norCode, isFlagSetting=True)
# NOT (vector)
notCode = 'destElem = ~srcElem1;'
sveUnaryInst('not', 'Not', 'SimdAluOp', unsignedTypes, notCode,
PredType.MERGE)
# ORN, ORNS (predicates)
ornCode = 'destElem = srcElem1 | !srcElem2;';
svePredLogicalInst('orn', 'PredOrn', 'SimdPredAluOp', ('uint8_t',),
ornCode)
svePredLogicalInst('orns', 'PredOrns', 'SimdPredAluOp', ('uint8_t',),
ornCode, isFlagSetting=True)
# ORR (immediate)
orCode = 'destElem = srcElem1 | srcElem2;'
sveWideImmInst('orr', 'OrrImm', 'SimdAluOp', ('uint64_t',), orCode)
# ORR (vectors, predicated)
sveBinInst('orr', 'OrrPred', 'SimdAluOp', unsignedTypes, orCode,
PredType.MERGE, True)
# ORR (vectors, unpredicated)
orCode = 'destElem = srcElem1 | srcElem2;'
sveBinInst('orr', 'OrrUnpred', 'SimdAluOp', ('uint64_t',), orCode)
# ORR, ORRS (predicates)
svePredLogicalInst('orr', 'PredOrr', 'SimdPredAluOp', ('uint8_t',), orCode)
svePredLogicalInst('orrs', 'PredOrrs', 'SimdPredAluOp', ('uint8_t',),
orCode, isFlagSetting=True)
# ORV
orvCode = 'destElem |= srcElem1;'
sveAssocReducInst('orv', 'Orv', 'SimdReduceAluOp', unsignedTypes,
orvCode, '0')
# PFALSE
pfalseCode = '''
PDest_ub[0] = 0;
destPred.reset();
'''
svePredUnaryWImplicitSrcInst('pfalse', 'Pfalse', 'SimdPredAluOp',
pfalseCode)
# PFIRST
svePFirstInst('pfirst', 'Pfirst', 'SimdPredAluOp')
# PNEXT
svePNextInst('pnext', 'Pnext', 'SimdPredAluOp', unsignedTypes)
# PTEST
svePredTestInst('ptest', 'Ptest', 'SimdPredAluOp')
# PTRUE
svePtrueInst('ptrue', 'Ptrue', 'SimdPredAluOp', unsignedTypes, False)
# PTRUES
svePtrueInst('ptrues', 'Ptrues', 'SimdPredAluOp', unsignedTypes, True)
# PUNPKHI
sveUnpackInst('punpkhi', 'Punpkhi', 'SimdPredAluOp', unsignedWideSDTypes,
unpackHalf = Unpack.High, regType = SrcRegType.Predicate)
# PUNPKLO
sveUnpackInst('punpklo', 'Punpklo', 'SimdPredAluOp', unsignedWideSDTypes,
unpackHalf = Unpack.Low, regType = SrcRegType.Predicate)
# RBIT
rbitCode = '''
destElem = reverseBits(srcElem1);'''
sveUnaryInst('rbit', 'Rbit', 'SimdAluOp', unsignedTypes, rbitCode,
predType=PredType.MERGE, srcRegType=SrcRegType.Vector)
# RDFFR (unpredicated)
rdffrUnpredCode = '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<uint8_t>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {
PDest_ub[i] = Ffr_ub[i];
}'''
svePredUnaryWImplicitSrcInst('rdffr', 'RdffrUnpred', 'SimdPredAluOp',
rdffrUnpredCode)
# RDFFR, RDFFRS (predicated)
rdffrPredCode = '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<uint8_t>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {
if (GpOp_ub[i]) {
PDest_ub[i] = Ffr_ub[i];
} else {
PDest_ub[i] = false;
}
}'''
svePredUnaryWImplicitSrcInst('rdffr', 'RdffrPred', 'SimdPredAluOp',
rdffrPredCode, PredType.ZERO, False)
svePredUnaryWImplicitSrcInst('rdffrs', 'RdffrsPred', 'SimdPredAluOp',
rdffrPredCode, PredType.ZERO, True)
# RDVL
rdvlCode = sveEnabledCheckCode + '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<uint8_t>(
xc->tcBase());
XDest = eCount * (int64_t) imm;
'''
rdvlIop = ArmInstObjParams('rdvl', 'SveRdvl', 'RegImmOp', rdvlCode, [])
header_output += RegImmOpDeclare.subst(rdvlIop)
decoder_output += RegImmOpConstructor.subst(rdvlIop)
exec_output += BasicExecute.subst(rdvlIop)
# REV (predicate)
sveReverseElementsInst('rev', 'Revp', 'SimdPredAluOp', unsignedTypes,
srcType = SrcRegType.Predicate)
# REV (vector)
sveReverseElementsInst('rev', 'Revv', 'SimdAluOp', unsignedTypes,
srcType = SrcRegType.Vector)
# REVB
revCode = '''
%(revtype)s* srcPtr = reinterpret_cast<%(revtype)s*>(&srcElem1);
%(revtype)s* dstPtr = reinterpret_cast<%(revtype)s*>(&destElem);
uint8_t subelements = sizeof(Element) / sizeof(%(revtype)s);
for(int i = 0; i < subelements; ++i) {
dstPtr[subelements - i - 1] = srcPtr[i];
}'''
sveUnaryInst('revb', 'Revb', 'SimdAluOp',
['uint16_t', 'uint32_t', 'uint64_t'],
revCode % {'revtype' : 'uint8_t'}, predType=PredType.MERGE,
srcRegType=SrcRegType.Vector, decoder='Generic')
# REVH
sveUnaryInst('revh', 'Revh', 'SimdAluOp', ['uint32_t', 'uint64_t'],
revCode % {'revtype' : 'uint16_t'}, predType=PredType.MERGE,
srcRegType=SrcRegType.Vector, decoder='Generic')
# REVW
sveUnaryInst('revw', 'Revw', 'SimdAluOp', ['uint64_t'],
revCode % {'revtype' : 'uint32_t'}, predType=PredType.MERGE,
srcRegType=SrcRegType.Vector, decoder='Generic')
# SABD
abdCode = '''
destElem = (srcElem1 > srcElem2) ? (srcElem1 - srcElem2) :
(srcElem2 - srcElem1);
'''
sveBinInst('sabd', 'Sabd', 'SimdAddOp', signedTypes, abdCode,
PredType.MERGE, True)
# SADDV
addvCode = 'destElem += srcElem1;'
sveWideningAssocReducInst('saddv', 'Saddv', 'SimdReduceAddOp',
['int8_t, int64_t', 'int16_t, int64_t', 'int32_t, int64_t'],
addvCode, '0')
# SCVTF
scvtfCode = fpOp % ('fplibFixedToFP<DElement>('
'sext<sizeof(SElement) * 8>(srcElem1), 0,'
' false, FPCRRounding(fpscr), fpscr)')
sveCvtInst('scvtf', 'ScvtfNarrow', 'SimdCvtOp',
('uint16_t, uint16_t',
'uint32_t, uint16_t',
'uint64_t, uint16_t',
'uint32_t, uint32_t',
'uint64_t, uint32_t',
'uint64_t, uint64_t'),
scvtfCode, CvtDir.Narrow)
sveCvtInst('scvtf', 'ScvtfWiden', 'SimdCvtOp', ('uint32_t, uint64_t',),
scvtfCode, CvtDir.Widen)
# SDIV
sdivCode = '''
constexpr Element ELEM_MIN = std::numeric_limits<Element>::min();
destElem = (srcElem2 == 0) ? 0 :
(srcElem2 == -1 && srcElem1 == ELEM_MIN) ? ELEM_MIN :
(srcElem1 / srcElem2);
'''
sveBinInst('sdiv', 'Sdiv', 'SimdDivOp', signedTypes, sdivCode,
PredType.MERGE, True)
# SDIVR
sdivrCode = '''
constexpr Element ELEM_MIN = std::numeric_limits<Element>::min();
destElem = (srcElem1 == 0) ? 0 :
(srcElem1 == -1 && srcElem2 == ELEM_MIN) ? ELEM_MIN :
(srcElem2 / srcElem1);
'''
sveBinInst('sdivr', 'Sdivr', 'SimdDivOp', signedTypes, sdivrCode,
PredType.MERGE, True)
# SDOT (indexed)
sveDotInst('sdot', 'Sdoti', 'SimdAluOp', ['int8_t, int32_t',
'int16_t, int64_t'], isIndexed = True)
# SDOT (vectors)
sveDotInst('sdot', 'Sdotv', 'SimdAluOp', ['int8_t, int32_t',
'int16_t, int64_t'], isIndexed = False)
# SEL (predicates)
selCode = 'destElem = srcElem1;'
svePredLogicalInst('sel', 'PredSel', 'SimdPredAluOp', ('uint8_t',),
selCode, PredType.SELECT)
# SEL (vectors)
sveBinInst('sel', 'Sel', 'SimdAluOp', unsignedTypes, selCode,
PredType.SELECT, False)
# SETFFR
setffrCode = '''
Ffr_ub[0] = true;
destPred.set();'''
svePredWriteFfrInst('setffr', 'Setffr', 'SimdPredAluOp', setffrCode, True)
# SMAX (immediate)
maxCode = 'destElem = (srcElem1 > srcElem2) ? srcElem1 : srcElem2;'
sveWideImmInst('smax', 'SmaxImm', 'SimdCmpOp', signedTypes, maxCode)
# SMAX (vectors)
sveBinInst('smax', 'Smax', 'SimdCmpOp', signedTypes, maxCode,
PredType.MERGE, True)
# SMAXV
maxvCode = '''
if (srcElem1 > destElem)
destElem = srcElem1;
'''
sveAssocReducInst('smaxv', 'Smaxv', 'SimdReduceCmpOp', signedTypes,
maxvCode, 'std::numeric_limits<Element>::min()')
# SMIN (immediate)
minCode = 'destElem = (srcElem1 < srcElem2) ? srcElem1 : srcElem2;'
sveWideImmInst('smin', 'SminImm', 'SimdCmpOp', signedTypes, minCode)
# SMIN (vectors)
sveBinInst('smin', 'Smin', 'SimdCmpOp', signedTypes, minCode,
PredType.MERGE, True)
# SMINV
minvCode = '''
if (srcElem1 < destElem)
destElem = srcElem1;
'''
sveAssocReducInst('sminv', 'Sminv', 'SimdReduceCmpOp', signedTypes,
minvCode, 'std::numeric_limits<Element>::max()')
# SMULH
exec_output += '''
template <class T>
T do_mulh(T srcElem1, T srcElem2)
{
return ((int64_t)srcElem1 * (int64_t)srcElem2) >> sizeof(T) * 8;
}
int64_t do_mulh(int64_t srcElem1, int64_t srcElem2)
{
uint64_t x = (uint64_t) llabs(srcElem1);
uint64_t y = (uint64_t) llabs(srcElem2);
uint64_t a = x >> 32;
uint64_t b = x & 0xFFFFFFFF;
uint64_t c = y >> 32;
uint64_t d = y & 0xFFFFFFFF;
uint64_t hi = a * c;
uint64_t lo = b * d;
hi += (a * d) >> 32;
uint64_t tmp = lo;
lo += ((a * d) & 0xFFFFFFFF) << 32;
if (lo < tmp)
hi++;
hi += (b * c) >> 32;
tmp = lo;
lo += ((b * c) & 0xFFFFFFFF) << 32;
if (lo < tmp)
hi++;
uint64_t destElem = hi;
if ((srcElem1 < 0) ^ (srcElem2 < 0)) {
uint64_t tmp = lo = ~lo;
destElem = ~hi;
if (++lo < tmp)
destElem++;
}
return destElem;
}
uint64_t do_mulh(uint64_t srcElem1, uint64_t srcElem2)
{
uint64_t x = srcElem1;
uint64_t y = srcElem2;
uint64_t a = x >> 32;
uint64_t b = x & 0xFFFFFFFF;
uint64_t c = y >> 32;
uint64_t d = y & 0xFFFFFFFF;
uint64_t hi = a * c;
uint64_t lo = b * d;
hi += (a * d) >> 32;
uint64_t tmp = lo;
lo += ((a * d) & 0xFFFFFFFF) << 32;
if (lo < tmp)
hi++;
hi += (b * c) >> 32;
tmp = lo;
lo += ((b * c) & 0xFFFFFFFF) << 32;
if (lo < tmp)
hi++;
return hi;
}'''
mulhCode = '''
destElem = do_mulh(srcElem1, srcElem2);'''
sveBinInst('smulh', 'Smulh', 'SimdMultOp', signedTypes, mulhCode,
PredType.MERGE, True)
# SPLICE
sveSpliceInst('splice', 'Splice', 'SimdAluOp', unsignedTypes)
# SQADD (immediate)
sqaddCode = '''
destElem = srcElem1 + srcElem2;
bool negDest = (destElem < 0);
bool negSrc1 = (srcElem1 < 0);
bool negSrc2 = (srcElem2 < 0);
if ((negDest != negSrc1) && (negSrc1 == negSrc2)) {
destElem = static_cast<Element>(
(Element)1 << (sizeof(Element) * 8 - 1)
);
if (negDest)
destElem -= 1;
}
'''
sveWideImmInst('sqadd', 'SqaddImm', 'SimdAddOp', signedTypes, sqaddCode)
# SQADD (vectors)
sveBinInst('sqadd', 'Sqadd', 'SimdAddOp', signedTypes, sqaddCode)
# SQDECB, SQDECH, SQDECW, SQDECD (scalar, 32-bit)
sqdecCode = '''
destElem = srcElem1 - (count * imm);
bool negDest = (destElem < 0);
bool negSrc = (srcElem1 < 0);
if (!negDest && negSrc) {
destElem = static_cast<%(dstType)s>(
(%(dstType)s)1 << (sizeof(%(dstType)s) * 8 - 1)
);
}
'''
sveElemCountInst('sqdec', 'Sqdec32', 'SimdAluOp', signedTypes,
sqdecCode%{'dstType':'int32_t'}, destType = DestType.Scalar,
dstIs32b = True)
# SQDECB, SQDECH, SQDECW, SQDECD (scalar, 64-bit)
sveElemCountInst('sqdec', 'Sqdec', 'SimdAluOp', signedTypes,
sqdecCode%{'dstType':'int64_t'}, destType = DestType.Scalar,
dstIs32b = False)
# SQDECH, SQDECW, SQDECD (vector)
sveElemCountInst('sqdec', 'Sqdecv', 'SimdAluOp', bigSignedTypes,
sqdecCode%{'dstType':'Element'}, destType = DestType.Vector,
dstIs32b = False)
# SQDECP (scalar, 32-bit)
sqdecpCode = '''
destElem = srcElem - count;
bool negDest = (destElem < 0);
bool negSrc = (srcElem < 0);
bool posCount = (count >= 0);
if ((negDest != negSrc) && (negSrc == posCount)) {
destElem = std::numeric_limits<%s>::min();
if (negDest)
destElem -= 1;
}
'''
sqdecp32Code = '''
int32_t srcElem = WDest;
int32_t destElem;''' + (sqdecpCode % 'int32_t') + '''
if (destElem < 0) {
XDest = static_cast<uint32_t>(destElem) | ~mask(32);
} else {
XDest = destElem;
}
'''
svePredCountInst('sqdecp', 'Sqdecp32', 'SimdAluOp', signedTypes,
sqdecp32Code, DestType.Scalar, SrcSize.Src32bit)
# SQDECP (scalar, 64-bit)
sqdecp64Code = '''
int64_t srcElem = XDest;
int64_t destElem;''' + (sqdecpCode % 'int64_t') + '''
XDest = destElem;
'''
svePredCountInst('sqdecp', 'Sqdecp64', 'SimdAluOp', signedTypes,
sqdecp64Code, DestType.Scalar, SrcSize.Src64bit)
# SQDECP (vector)
svePredCountInst('sqdecp', 'Sqdecpv', 'SimdAluOp', signedTypes,
sqdecpCode % 'Element', DestType.Vector)
# SQINCB, SQINCH, SQINCW, SQINCD (scalar, 32-bit)
sqincCode = '''
destElem = srcElem1 + (count * imm);
bool negDest = (destElem < 0);
bool negSrc = (srcElem1 < 0);
if (negDest && !negSrc) {
destElem = static_cast<%(dstType)s>(
(%(dstType)s)1 << (sizeof(%(dstType)s) * 8 - 1)
);
destElem -= 1;
}
'''
sveElemCountInst('sqinc', 'Sqinc32', 'SimdAluOp', signedTypes,
sqincCode%{'dstType':'int32_t'}, destType = DestType.Scalar,
dstIs32b = True)
# SQINCB, SQINCH, SQINCW, SQINCD (scalar, 64-bit)
sveElemCountInst('sqinc', 'Sqinc', 'SimdAluOp', signedTypes,
sqincCode%{'dstType':'int64_t'}, destType = DestType.Scalar,
dstIs32b = False)
# SQINCH, SQINCW, SQINCD (vector)
sveElemCountInst('sqinc', 'Sqincv', 'SimdAluOp', bigSignedTypes,
sqincCode%{'dstType':'Element'}, destType = DestType.Vector,
dstIs32b = False)
# SQINCP (scalar, 32-bit)
sqincpCode = '''
destElem = srcElem + count;
bool negDest = (destElem < 0);
bool negSrc = (srcElem < 0);
bool negCount = (count < 0);
if ((negDest != negSrc) && (negSrc == negCount)) {
destElem = std::numeric_limits<%s>::min();
if (negDest)
destElem -= 1;
}
'''
sqincp32Code = '''
int32_t srcElem = WDest;
int32_t destElem;''' + (sqincpCode % 'int32_t') + '''
if (destElem < 0) {
XDest = static_cast<uint32_t>(destElem) | ~mask(32);
} else {
XDest = destElem;
}
'''
svePredCountInst('sqincp', 'Sqincp32', 'SimdAluOp', signedTypes,
sqincp32Code, DestType.Scalar, SrcSize.Src32bit)
# SQINCP (scalar, 64-bit)
sqincp64Code = '''
int64_t srcElem = XDest;
int64_t destElem;''' + (sqincpCode % 'int64_t') + '''
XDest = destElem;
'''
svePredCountInst('sqincp', 'Sqincp64', 'SimdAluOp', signedTypes,
sqincp64Code, DestType.Scalar, SrcSize.Src64bit)
# SQINCP (vector)
svePredCountInst('sqincp', 'Sqincpv', 'SimdAluOp', signedTypes,
sqincpCode % 'Element', DestType.Vector)
# SQSUB (immediate)
sqsubCode = '''
destElem = srcElem1 - srcElem2;
bool negDest = (destElem < 0);
bool negSrc1 = (srcElem1 < 0);
bool posSrc2 = (srcElem2 >= 0);
if ((negDest != negSrc1) && (negSrc1 == posSrc2)) {
destElem = static_cast<Element>(
(Element)1 << (sizeof(Element) * 8 - 1)
);
if (negDest)
destElem -= 1;
}
'''
sveWideImmInst('sqsub', 'SqsubImm', 'SimdAddOp', signedTypes, sqsubCode)
# SQSUB (vectors)
sveBinInst('sqsub', 'Sqsub', 'SimdAddOp', signedTypes, sqsubCode)
# SUB (immediate)
subCode = 'destElem = srcElem1 - srcElem2;'
sveWideImmInst('sub', 'SubImm', 'SimdAddOp', unsignedTypes, subCode)
# SUB (vectors, predicated)
sveBinInst('sub', 'SubPred', 'SimdAddOp', unsignedTypes, subCode,
PredType.MERGE, True)
# SUB (vectors, unpredicated)
subCode = 'destElem = srcElem1 - srcElem2;'
sveBinInst('sub', 'SubUnpred', 'SimdAddOp', unsignedTypes, subCode)
# SUBR (immediate)
subrCode = 'destElem = srcElem2 - srcElem1;'
sveWideImmInst('subr', 'SubrImm', 'SimdAddOp', unsignedTypes, subrCode)
# SUBR (vectors)
sveBinInst('subr', 'Subr', 'SimdAddOp', unsignedTypes, subrCode,
PredType.MERGE, True)
# SUNPKHI
sveUnpackInst('sunpkhi', 'Sunpkhi', 'SimdAluOp', signedWideSDTypes,
unpackHalf = Unpack.High, regType = SrcRegType.Vector)
# SUNPKLO
sveUnpackInst('sunpklo', 'Sunpklo', 'SimdAluOp', signedWideSDTypes,
unpackHalf = Unpack.Low, regType = SrcRegType.Vector)
# SXTB
sxtCode = 'destElem = sext<8 * sizeof(SElement)>(srcElem1);'
sveWidenUnaryInst('sxtb', 'Sxtb', 'SimdAluOp',
['uint8_t, uint16_t', 'uint8_t, uint32_t', 'uint8_t, uint64_t'],
sxtCode, PredType.MERGE)
# SXTH
sveWidenUnaryInst('sxth', 'Sxth', 'SimdAluOp',
['uint16_t, uint32_t', 'uint16_t, uint64_t'],
sxtCode, PredType.MERGE)
# SXTW
sveWidenUnaryInst('sxtw', 'Sxtw', 'SimdAluOp',
['uint32_t, uint64_t'],
sxtCode, PredType.MERGE)
# TBL
sveTblInst('tbl', 'Tbl', 'SimdAluOp')
# TRN1, TRN2 (predicates)
trnPredIterCode = '''
constexpr unsigned sz = sizeof(Element);
int s;
int part = %d;
ArmISA::VecPredRegContainer tmpPredC;
auto auxPDest = tmpPredC.as<uint8_t>();
for (unsigned i = 0; i < eCount / 2; i++) {
s = 2 * i + part;
for (unsigned j = 0; j < sz; j++) {
auxPDest[(2 * i) * sz + j] = POp1_pb[s * sz + j];
auxPDest[(2 * i + 1) * sz + j] = POp2_pb[s * sz + j];
}
}
for (unsigned i = 0; i < eCount * sz; i++) {
PDest_pb[i] = auxPDest[i];
}
'''
svePredBinPermInst('trn1', 'Trn1Pred', 'SimdPredAluOp', unsignedTypes,
trnPredIterCode % 0)
svePredBinPermInst('trn2', 'Trn2Pred', 'SimdPredAluOp', unsignedTypes,
trnPredIterCode % 1)
# TRN1, TRN2 (vectors)
trnIterCode = '''
int s;
int part = %d;
ArmISA::VecRegContainer tmpVecC;
auto auxDest = tmpVecC.as<Element>();
for (unsigned i = 0; i < eCount / 2; i++) {
s = 2 * i + part;
auxDest[2 * i] = AA64FpOp1_x[s];
auxDest[2 * i + 1] = AA64FpOp2_x[s];
}
for (unsigned i = 0; i < eCount; i++) {
AA64FpDest_x[i] = auxDest[i];
}
'''
sveBinInst('trn1', 'Trn1', 'SimdAluOp', unsignedTypes, '',
customIterCode=trnIterCode % 0)
sveBinInst('trn2', 'Trn2', 'SimdAluOp', unsignedTypes, '',
customIterCode=trnIterCode % 1)
# UABD
sveBinInst('uabd', 'Uabd', 'SimdAddOp', unsignedTypes, abdCode,
PredType.MERGE, True)
# UADDV
sveWideningAssocReducInst('uaddv', 'Uaddv', 'SimdReduceAddOp',
['uint8_t, uint64_t', 'uint16_t, uint64_t', 'uint32_t, uint64_t',
'uint64_t, uint64_t'],
addvCode, '0')
# UCVTF
ucvtfCode = fpOp % ('fplibFixedToFP<DElement>(srcElem1, 0, true,'
' FPCRRounding(fpscr), fpscr)')
sveCvtInst('ucvtf', 'UcvtfNarrow', 'SimdCvtOp',
('uint16_t, uint16_t',
'uint32_t, uint16_t',
'uint64_t, uint16_t',
'uint32_t, uint32_t',
'uint64_t, uint32_t',
'uint64_t, uint64_t'),
ucvtfCode, CvtDir.Narrow)
sveCvtInst('ucvtf', 'UcvtfWiden', 'SimdCvtOp', ('uint32_t, uint64_t',),
ucvtfCode, CvtDir.Widen)
# UDIV
udivCode = 'destElem = (srcElem2 == 0) ? 0 : (srcElem1 / srcElem2);'
sveBinInst('udiv', 'Udiv', 'SimdDivOp', unsignedTypes, udivCode,
PredType.MERGE, True)
# UDIVR
udivrCode = 'destElem = (srcElem1 == 0) ? 0 : (srcElem2 / srcElem1);'
sveBinInst('udivr', 'Udivr', 'SimdDivOp', unsignedTypes, udivrCode,
PredType.MERGE, True)
# UDOT (indexed)
sveDotInst('udot', 'Udoti', 'SimdAluOp', ['uint8_t, uint32_t',
'uint16_t, uint64_t'], isIndexed = True)
# UDOT (vectors)
sveDotInst('udot', 'Udotv', 'SimdAluOp', ['uint8_t, uint32_t',
'uint16_t, uint64_t'], isIndexed = False)
# UMAX (immediate)
sveWideImmInst('umax', 'UmaxImm', 'SimdCmpOp', unsignedTypes, maxCode)
# UMAX (vectors)
sveBinInst('umax', 'Umax', 'SimdCmpOp', unsignedTypes, maxCode,
PredType.MERGE, True)
# UMAXV
sveAssocReducInst('umaxv', 'Umaxv', 'SimdReduceCmpOp', unsignedTypes,
maxvCode, 'std::numeric_limits<Element>::min()')
# UMIN (immediate)
sveWideImmInst('umin', 'UminImm', 'SimdCmpOp', unsignedTypes, minCode)
# UMIN (vectors)
sveBinInst('umin', 'Umin', 'SimdCmpOp', unsignedTypes, minCode,
PredType.MERGE, True)
# UMINV
sveAssocReducInst('uminv', 'Uminv', 'SimdReduceCmpOp', unsignedTypes,
minvCode, 'std::numeric_limits<Element>::max()')
# UMULH
sveBinInst('umulh', 'Umulh', 'SimdMultOp', unsignedTypes, mulhCode,
PredType.MERGE, True)
# UQADD (immediate)
uqaddCode = '''
destElem = srcElem1 + srcElem2;
if (destElem < srcElem1 || destElem < srcElem2) {
destElem = (Element)(-1);
}
'''
sveWideImmInst('uqadd', 'UqaddImm', 'SimdAddOp', unsignedTypes, uqaddCode)
# UQADD (vectors)
sveBinInst('uqadd', 'Uqadd', 'SimdAddOp', unsignedTypes, uqaddCode)
# UQDECB, UQDECH, UQDECW, UQDECD (scalar, 32-bit)
uqdecCode = '''
destElem = srcElem1 - (imm * count);
if (destElem > srcElem1) {
destElem = 0;
}
'''
sveElemCountInst('uqdec', 'Uqdec32', 'SimdAluOp', unsignedTypes,
uqdecCode, destType = DestType.Scalar, dstIs32b = True)
# UQDECB, UQDECH, UQDECW, UQDECD (scalar, 64-bit)
sveElemCountInst('uqdec', 'Uqdec', 'SimdAluOp', unsignedTypes,
uqdecCode, destType = DestType.Scalar, dstIs32b = False)
# UQDECH, UQDECW, UQDECD (vector)
sveElemCountInst('uqdec', 'Uqdecv', 'SimdAluOp', bigUnsignedTypes,
uqdecCode, destType = DestType.Vector, dstIs32b = False)
# UQDECP (scalar, 32-bit)
uqdecpCode = '''
destElem = srcElem - count;
if (destElem > srcElem) {
destElem = 0;
}
'''
uqdecp32Code = '''
uint32_t srcElem = WDest;
uint32_t destElem;''' + uqdecpCode + '''
WDest = destElem;
'''
svePredCountInst('uqdecp', 'Uqdecp32', 'SimdAluOp', unsignedTypes,
uqdecp32Code, DestType.Scalar, SrcSize.Src32bit)
# UQDECP (scalar, 64-bit)
uqdecp64Code = '''
uint64_t srcElem = XDest;
uint64_t destElem;''' + uqdecpCode + '''
XDest = destElem;
'''
svePredCountInst('uqdecp', 'Uqdecp64', 'SimdAluOp', unsignedTypes,
uqdecp64Code, DestType.Scalar, SrcSize.Src64bit)
# UQDECP (vector)
svePredCountInst('uqdecp', 'Uqdecpv', 'SimdAluOp', unsignedTypes,
uqdecpCode, DestType.Vector)
# UQDECB, UQDECH, UQDECW, UQDECD (scalar, 32-bit)
uqincCode = '''
destElem = srcElem1 + (imm * count);
if (destElem < srcElem1 || destElem < (imm * count)) {
destElem = static_cast<%(destType)s>(-1);
}
'''
sveElemCountInst('uqinc', 'Uqinc32', 'SimdAluOp', unsignedTypes,
uqincCode%{'destType': 'uint32_t'}, destType = DestType.Scalar,
dstIs32b = True)
# UQDECB, UQDECH, UQDECW, UQDECD (scalar, 64-bit)
sveElemCountInst('uqinc', 'Uqinc', 'SimdAluOp', unsignedTypes,
uqincCode%{'destType': 'uint64_t'}, destType = DestType.Scalar,
dstIs32b = False)
# UQDECH, UQDECW, UQDECD (vector)
sveElemCountInst('uqinc', 'Uqincv', 'SimdAluOp', bigUnsignedTypes,
uqincCode%{'destType': 'Element'}, destType = DestType.Vector,
dstIs32b = False)
# UQINCP (scalar, 32-bit)
uqincpCode = '''
destElem = srcElem + count;
if (destElem < srcElem || destElem < count) {
destElem = std::numeric_limits<%s>::max();
}
'''
uqincp32Code = '''
uint32_t srcElem = WDest;
uint32_t destElem;''' + (uqincpCode % 'uint32_t') + '''
XDest = destElem;
'''
svePredCountInst('uqincp', 'Uqincp32', 'SimdAluOp', unsignedTypes,
uqincp32Code, DestType.Scalar, SrcSize.Src32bit)
# UQINCP (scalar, 64-bit)
uqincp64Code = '''
uint64_t srcElem = XDest;
uint64_t destElem;''' + (uqincpCode % 'uint64_t') + '''
XDest = destElem;
'''
svePredCountInst('uqincp', 'Uqincp64', 'SimdAluOp', unsignedTypes,
uqincp64Code, DestType.Scalar, SrcSize.Src64bit)
# UQINCP (vector)
svePredCountInst('uqincp', 'Uqincpv', 'SimdAluOp', unsignedTypes,
uqincpCode % 'Element', DestType.Vector)
# UQSUB (immediate)
uqsubCode = '''
destElem = srcElem1 - srcElem2;
if (destElem > srcElem1) {
destElem = 0;
}
'''
sveWideImmInst('uqsub', 'UqsubImm', 'SimdAddOp', unsignedTypes, uqsubCode)
# UQSUB (vectors)
sveBinInst('uqsub', 'Uqsub', 'SimdAddOp', unsignedTypes, uqsubCode)
# UUNPKHI
sveUnpackInst('uunpkhi', 'Uunpkhi', 'SimdAluOp', unsignedWideSDTypes,
unpackHalf = Unpack.High, regType = SrcRegType.Vector)
# UUNPKLO
sveUnpackInst('uunpklo', 'Uunpklo', 'SimdAluOp', unsignedWideSDTypes,
unpackHalf = Unpack.Low, regType = SrcRegType.Vector)
# UXTB
uxtCode = 'destElem = srcElem1;'
sveWidenUnaryInst('uxtb', 'Uxtb', 'SimdAluOp',
['uint8_t, uint16_t', 'uint8_t, uint32_t', 'uint8_t, uint64_t'],
uxtCode, PredType.MERGE)
# UXTH
sveWidenUnaryInst('uxth', 'Uxth', 'SimdAluOp',
['uint16_t, uint32_t', 'uint16_t, uint64_t'],
uxtCode, PredType.MERGE)
# UXTW
sveWidenUnaryInst('uxtw', 'Uxtw', 'SimdAluOp',
['uint32_t, uint64_t'],
uxtCode, PredType.MERGE)
# UZP1, UZP2 (predicates)
uzpPredIterCode = '''
constexpr unsigned sz = sizeof(Element);
int s;
int part = %d;
ArmISA::VecPredRegContainer tmpPredC;
auto auxPDest = tmpPredC.as<uint8_t>();
for (unsigned i = 0; i < eCount; i++) {
s = 2 * i + part;
for (unsigned j = 0; j < sz; j++) {
if (s < eCount) {
auxPDest[i * sz + j] = POp1_pb[s * sz + j];
} else {
auxPDest[i * sz + j] = POp2_pb[(s - eCount) * sz + j];
}
}
}
for (unsigned i = 0; i < eCount * sz; i++) {
PDest_pb[i] = auxPDest[i];
}
'''
svePredBinPermInst('uzp1', 'Uzp1Pred', 'SimdPredAluOp', unsignedTypes,
uzpPredIterCode % 0)
svePredBinPermInst('uzp2', 'Uzp2Pred', 'SimdPredAluOp', unsignedTypes,
uzpPredIterCode % 1)
# UZP1, UZP2 (vectors)
uzpIterCode = '''
int s;
int part = %d;
ArmISA::VecRegContainer tmpVecC;
auto auxDest = tmpVecC.as<Element>();
for (unsigned i = 0; i < eCount; i++) {
s = 2 * i + part;
if (s < eCount) {
auxDest[i] = AA64FpOp1_x[s];
} else {
auxDest[i] = AA64FpOp2_x[s - eCount];
}
}
for (unsigned i = 0; i < eCount; i++) {
AA64FpDest_x[i] = auxDest[i];
}
'''
sveBinInst('uzp1', 'Uzp1', 'SimdAluOp', unsignedTypes, '',
customIterCode=uzpIterCode % 0)
sveBinInst('uzp2', 'Uzp2', 'SimdAluOp', unsignedTypes, '',
customIterCode=uzpIterCode % 1)
# WHILELE (32-bit)
whileLECode = '''
cond = srcElem1 <= srcElem2;
'''
sveWhileInst('whilele', 'Whilele32', 'SimdCmpOp', signedTypes, whileLECode,
SrcSize.Src32bit)
# WHILELE (64-bit)
sveWhileInst('whilele', 'Whilele64', 'SimdCmpOp', signedTypes, whileLECode,
SrcSize.Src64bit)
# WHILELO (32-bit)
whileLTCode = '''
cond = srcElem1 < srcElem2;
'''
sveWhileInst('whilelo', 'Whilelo32', 'SimdCmpOp', unsignedTypes,
whileLTCode, SrcSize.Src32bit)
# WHILELO (64-bit)
sveWhileInst('whilelo', 'Whilelo64', 'SimdCmpOp', unsignedTypes,
whileLTCode, SrcSize.Src64bit)
# WHILELS (32-bit)
sveWhileInst('whilels', 'Whilels32', 'SimdCmpOp', unsignedTypes,
whileLECode, SrcSize.Src32bit)
# WHILELS (64-bit)
sveWhileInst('whilels', 'Whilels64', 'SimdCmpOp', unsignedTypes,
whileLECode, SrcSize.Src64bit)
# WHILELT (32-bit)
sveWhileInst('whilelt', 'Whilelt32', 'SimdCmpOp', signedTypes,
whileLTCode, SrcSize.Src32bit)
# WHILELT (64-bit)
sveWhileInst('whilelt', 'Whilelt64', 'SimdCmpOp', signedTypes,
whileLTCode, SrcSize.Src64bit)
# WRFFR
wrffrCode = '''
unsigned eCount = ArmStaticInst::getCurSveVecLen<uint8_t>(
xc->tcBase());
for (unsigned i = 0; i < eCount; i++) {
Ffr_ub[i] = POp1_ub[i];
}'''
svePredWriteFfrInst('wrffr', 'Wrffr', 'SimdPredAluOp', wrffrCode, False)
# ZIP1, ZIP2 (predicates)
zipPredIterCode = '''
constexpr unsigned sz = sizeof(Element);
int s;
int part = %d;
ArmISA::VecPredRegContainer tmpPredC;
auto auxPDest = tmpPredC.as<uint8_t>();
for (unsigned i = 0; i < eCount / 2; i++) {
s = i + (part * (eCount / 2));
for (unsigned j = 0; j < sz; j++) {
auxPDest[(2 * i) * sz + j] = POp1_pb[s * sz + j];
auxPDest[(2 * i + 1) * sz + j] = POp2_pb[s * sz + j];
}
}
for (unsigned i = 0; i < eCount * sz; i++) {
PDest_pb[i] = auxPDest[i];
}
'''
svePredBinPermInst('zip1', 'Zip1Pred', 'SimdPredAluOp', unsignedTypes,
zipPredIterCode % 0)
svePredBinPermInst('zip2', 'Zip2Pred', 'SimdPredAluOp', unsignedTypes,
zipPredIterCode % 1)
# ZIP1, ZIP2 (vectors)
zipIterCode = '''
int s;
int part = %d;
ArmISA::VecRegContainer tmpVecC;
auto auxDest = tmpVecC.as<Element>();
for (unsigned i = 0; i < eCount / 2; i++) {
s = i + (part * (eCount / 2));
auxDest[2 * i] = AA64FpOp1_x[s];
auxDest[2 * i + 1] = AA64FpOp2_x[s];
}
for (unsigned i = 0; i < eCount; i++) {
AA64FpDest_x[i] = auxDest[i];
}
'''
sveBinInst('zip1', 'Zip1', 'SimdAluOp', unsignedTypes, '',
customIterCode=zipIterCode % 0)
sveBinInst('zip2', 'Zip2', 'SimdAluOp', unsignedTypes, '',
customIterCode=zipIterCode % 1)
}};