| // Copyright (c) 2017-2018 ARM Limited |
| // All rights reserved |
| // |
| // The license below extends only to copyright in the software and shall |
| // not be construed as granting a license to any other intellectual |
| // property including but not limited to intellectual property relating |
| // to a hardware implementation of the functionality of the software |
| // licensed hereunder. You may use the software subject to the license |
| // terms below provided that you ensure that this notice is replicated |
| // unmodified and in its entirety in all distributions of the software, |
| // modified or unmodified, in source code or in binary form. |
| // |
| // Redistribution and use in source and binary forms, with or without |
| // modification, are permitted provided that the following conditions are |
| // met: redistributions of source code must retain the above copyright |
| // notice, this list of conditions and the following disclaimer; |
| // redistributions in binary form must reproduce the above copyright |
| // notice, this list of conditions and the following disclaimer in the |
| // documentation and/or other materials provided with the distribution; |
| // neither the name of the copyright holders nor the names of its |
| // contributors may be used to endorse or promote products derived from |
| // this software without specific prior written permission. |
| // |
| // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| // |
| // Authors: Giacomo Gabrielli |
| |
| // @file Definition of SVE memory access instructions. |
| |
| output header {{ |
| |
| // Decodes SVE contiguous load instructions, scalar plus scalar form. |
| template <template <typename T1, typename T2> class Base> |
| StaticInstPtr |
| decodeSveContigLoadSSInsts(uint8_t dtype, ExtMachInst machInst, |
| IntRegIndex zt, IntRegIndex pg, IntRegIndex rn, |
| IntRegIndex rm, bool firstFaulting) |
| { |
| const char* mn = firstFaulting ? "ldff1" : "ld1"; |
| switch (dtype) { |
| case 0x0: |
| return new Base<uint8_t, uint8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x1: |
| return new Base<uint16_t, uint8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x2: |
| return new Base<uint32_t, uint8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x3: |
| return new Base<uint64_t, uint8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x4: |
| return new Base<int64_t, int32_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x5: |
| return new Base<uint16_t, uint16_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x6: |
| return new Base<uint32_t, uint16_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x7: |
| return new Base<uint64_t, uint16_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x8: |
| return new Base<int64_t, int16_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x9: |
| return new Base<int32_t, int16_t>(mn, machInst, zt, pg, rn, rm); |
| case 0xa: |
| return new Base<uint32_t, uint32_t>(mn, machInst, zt, pg, rn, rm); |
| case 0xb: |
| return new Base<uint64_t, uint32_t>(mn, machInst, zt, pg, rn, rm); |
| case 0xc: |
| return new Base<int64_t, int8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0xd: |
| return new Base<int32_t, int8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0xe: |
| return new Base<int16_t, int8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0xf: |
| return new Base<uint64_t, uint64_t>(mn, machInst, zt, pg, rn, rm); |
| } |
| return new Unknown64(machInst); |
| } |
| |
| // Decodes SVE contiguous load instructions, scalar plus immediate form. |
| template <template <typename T1, typename T2> class Base> |
| StaticInstPtr |
| decodeSveContigLoadSIInsts(uint8_t dtype, ExtMachInst machInst, |
| IntRegIndex zt, IntRegIndex pg, IntRegIndex rn, |
| uint64_t imm, bool nonFaulting, |
| bool replicate = false) |
| { |
| assert(!(nonFaulting && replicate)); |
| const char* mn = replicate ? "ld1r" : (nonFaulting ? "ldnf1" : "ld1"); |
| switch (dtype) { |
| case 0x0: |
| return new Base<uint8_t, uint8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x1: |
| return new Base<uint16_t, uint8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x2: |
| return new Base<uint32_t, uint8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x3: |
| return new Base<uint64_t, uint8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x4: |
| return new Base<int64_t, int32_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x5: |
| return new Base<uint16_t, uint16_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x6: |
| return new Base<uint32_t, uint16_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x7: |
| return new Base<uint64_t, uint16_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x8: |
| return new Base<int64_t, int16_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x9: |
| return new Base<int32_t, int16_t>(mn, machInst, zt, pg, rn, imm); |
| case 0xa: |
| return new Base<uint32_t, uint32_t>(mn, machInst, zt, pg, rn, imm); |
| case 0xb: |
| return new Base<uint64_t, uint32_t>(mn, machInst, zt, pg, rn, imm); |
| case 0xc: |
| return new Base<int64_t, int8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0xd: |
| return new Base<int32_t, int8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0xe: |
| return new Base<int16_t, int8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0xf: |
| return new Base<uint64_t, uint64_t>(mn, machInst, zt, pg, rn, imm); |
| } |
| return new Unknown64(machInst); |
| } |
| |
| // Decodes SVE contiguous store instructions, scalar plus scalar form. |
| template <template <typename T1, typename T2> class Base> |
| StaticInstPtr |
| decodeSveContigStoreSSInsts(uint8_t dtype, ExtMachInst machInst, |
| IntRegIndex zt, IntRegIndex pg, IntRegIndex rn, |
| IntRegIndex rm) |
| { |
| const char* mn = "st1"; |
| switch (dtype) { |
| case 0x0: |
| return new Base<uint8_t, uint8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x1: |
| return new Base<uint16_t, uint8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x2: |
| return new Base<uint32_t, uint8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x3: |
| return new Base<uint64_t, uint8_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x5: |
| return new Base<uint16_t, uint16_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x6: |
| return new Base<uint32_t, uint16_t>(mn, machInst, zt, pg, rn, rm); |
| case 0x7: |
| return new Base<uint64_t, uint16_t>(mn, machInst, zt, pg, rn, rm); |
| case 0xa: |
| return new Base<uint32_t, uint32_t>(mn, machInst, zt, pg, rn, rm); |
| case 0xb: |
| return new Base<uint64_t, uint32_t>(mn, machInst, zt, pg, rn, rm); |
| case 0xf: |
| return new Base<uint64_t, uint64_t>(mn, machInst, zt, pg, rn, rm); |
| } |
| return new Unknown64(machInst); |
| } |
| |
| // Decodes SVE contiguous store instructions, scalar plus immediate form. |
| template <template <typename T1, typename T2> class Base> |
| StaticInstPtr |
| decodeSveContigStoreSIInsts(uint8_t dtype, ExtMachInst machInst, |
| IntRegIndex zt, IntRegIndex pg, IntRegIndex rn, |
| int8_t imm) |
| { |
| const char* mn = "st1"; |
| switch (dtype) { |
| case 0x0: |
| return new Base<uint8_t, uint8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x1: |
| return new Base<uint16_t, uint8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x2: |
| return new Base<uint32_t, uint8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x3: |
| return new Base<uint64_t, uint8_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x5: |
| return new Base<uint16_t, uint16_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x6: |
| return new Base<uint32_t, uint16_t>(mn, machInst, zt, pg, rn, imm); |
| case 0x7: |
| return new Base<uint64_t, uint16_t>(mn, machInst, zt, pg, rn, imm); |
| case 0xa: |
| return new Base<uint32_t, uint32_t>(mn, machInst, zt, pg, rn, imm); |
| case 0xb: |
| return new Base<uint64_t, uint32_t>(mn, machInst, zt, pg, rn, imm); |
| case 0xf: |
| return new Base<uint64_t, uint64_t>(mn, machInst, zt, pg, rn, imm); |
| } |
| return new Unknown64(machInst); |
| } |
| |
| // NOTE: SVE load-and-replicate instructions are decoded with |
| // decodeSveContigLoadSIInsts(...). |
| |
| }}; |
| |
| output decoder {{ |
| |
| StaticInstPtr |
| decodeSveGatherLoadVIInsts(uint8_t dtype, ExtMachInst machInst, |
| IntRegIndex zt, IntRegIndex pg, IntRegIndex zn, |
| uint64_t imm, bool esizeIs32, |
| bool firstFault) |
| { |
| const char* mn = firstFault ? "ldff1" : "ld1"; |
| switch (dtype) { |
| case 0x0: |
| if (esizeIs32) { |
| return new SveIndexedMemVI<int32_t, int8_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } else { |
| return new SveIndexedMemVI<int64_t, int8_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } |
| case 0x1: |
| if (esizeIs32) { |
| return new SveIndexedMemVI<uint32_t, uint8_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } else { |
| return new SveIndexedMemVI<uint64_t, uint8_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } |
| case 0x2: |
| if (esizeIs32) { |
| return new SveIndexedMemVI<int32_t, int16_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } else { |
| return new SveIndexedMemVI<int64_t, int16_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } |
| case 0x3: |
| if (esizeIs32) { |
| return new SveIndexedMemVI<uint32_t, uint16_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } else { |
| return new SveIndexedMemVI<uint64_t, uint16_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } |
| case 0x4: |
| if (esizeIs32) { |
| break; |
| } else { |
| return new SveIndexedMemVI<int64_t, int32_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } |
| case 0x5: |
| if (esizeIs32) { |
| return new SveIndexedMemVI<uint32_t, uint32_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } else { |
| return new SveIndexedMemVI<uint64_t, uint32_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } |
| case 0x7: |
| if (esizeIs32) { |
| break; |
| } else { |
| return new SveIndexedMemVI<uint64_t, uint64_t, |
| SveGatherLoadVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, zn, imm, firstFault); |
| } |
| } |
| return new Unknown64(machInst); |
| } |
| |
| StaticInstPtr |
| decodeSveGatherLoadSVInsts(uint8_t dtype, ExtMachInst machInst, |
| IntRegIndex zt, IntRegIndex pg, IntRegIndex rn, |
| IntRegIndex zm, bool esizeIs32, bool offsetIs32, |
| bool offsetIsSigned, bool offsetIsScaled, |
| bool firstFault) |
| { |
| const char* mn = firstFault ? "ldff1" : "ld1"; |
| switch (dtype) { |
| case 0x0: |
| if (esizeIs32) { |
| return new SveIndexedMemSV<int32_t, int8_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } else { |
| return new SveIndexedMemSV<int64_t, int8_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } |
| case 0x1: |
| if (esizeIs32) { |
| return new SveIndexedMemSV<uint32_t, uint8_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } else { |
| return new SveIndexedMemSV<uint64_t, uint8_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } |
| case 0x2: |
| if (esizeIs32) { |
| return new SveIndexedMemSV<int32_t, int16_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } else { |
| return new SveIndexedMemSV<int64_t, int16_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } |
| case 0x3: |
| if (esizeIs32) { |
| return new SveIndexedMemSV<uint32_t, uint16_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } else { |
| return new SveIndexedMemSV<uint64_t, uint16_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } |
| case 0x4: |
| if (esizeIs32) { |
| break; |
| } else { |
| return new SveIndexedMemSV<int64_t, int32_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } |
| case 0x5: |
| if (esizeIs32) { |
| return new SveIndexedMemSV<uint32_t, uint32_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } else { |
| return new SveIndexedMemSV<uint64_t, uint32_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } |
| case 0x7: |
| if (esizeIs32) { |
| break; |
| } else { |
| return new SveIndexedMemSV<uint64_t, uint64_t, |
| SveGatherLoadSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemReadOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, firstFault); |
| } |
| } |
| return new Unknown64(machInst); |
| } |
| |
| StaticInstPtr |
| decodeSveScatterStoreVIInsts(uint8_t msz, ExtMachInst machInst, |
| IntRegIndex zt, IntRegIndex pg, |
| IntRegIndex zn, uint64_t imm, |
| bool esizeIs32) |
| { |
| const char* mn = "st1"; |
| switch (msz) { |
| case 0x0: |
| if (esizeIs32) { |
| return new SveIndexedMemVI<uint32_t, uint8_t, |
| SveScatterStoreVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, zn, imm, false); |
| } else { |
| return new SveIndexedMemVI<uint64_t, uint8_t, |
| SveScatterStoreVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, zn, imm, false); |
| } |
| case 0x1: |
| if (esizeIs32) { |
| return new SveIndexedMemVI<uint32_t, uint16_t, |
| SveScatterStoreVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, zn, imm, false); |
| } else { |
| return new SveIndexedMemVI<uint64_t, uint16_t, |
| SveScatterStoreVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, zn, imm, false); |
| } |
| case 0x2: |
| if (esizeIs32) { |
| return new SveIndexedMemVI<uint32_t, uint32_t, |
| SveScatterStoreVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, zn, imm, false); |
| } else { |
| return new SveIndexedMemVI<uint64_t, uint32_t, |
| SveScatterStoreVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, zn, imm, false); |
| } |
| case 0x3: |
| if (esizeIs32) { |
| break; |
| } else { |
| return new SveIndexedMemVI<uint64_t, uint64_t, |
| SveScatterStoreVIMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, zn, imm, false); |
| } |
| } |
| return new Unknown64(machInst); |
| } |
| |
| StaticInstPtr |
| decodeSveScatterStoreSVInsts(uint8_t msz, ExtMachInst machInst, |
| IntRegIndex zt, IntRegIndex pg, |
| IntRegIndex rn, IntRegIndex zm, |
| bool esizeIs32, bool offsetIs32, |
| bool offsetIsSigned, bool offsetIsScaled) |
| { |
| const char* mn = "st1"; |
| switch (msz) { |
| case 0x0: |
| if (esizeIs32) { |
| return new SveIndexedMemSV<uint32_t, uint8_t, |
| SveScatterStoreSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, false); |
| } else { |
| return new SveIndexedMemSV<uint64_t, uint8_t, |
| SveScatterStoreSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, false); |
| } |
| case 0x1: |
| if (esizeIs32) { |
| return new SveIndexedMemSV<uint32_t, uint16_t, |
| SveScatterStoreSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, false); |
| } else { |
| return new SveIndexedMemSV<uint64_t, uint16_t, |
| SveScatterStoreSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, false); |
| } |
| case 0x2: |
| if (esizeIs32) { |
| return new SveIndexedMemSV<uint32_t, uint32_t, |
| SveScatterStoreSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, false); |
| } else { |
| return new SveIndexedMemSV<uint64_t, uint32_t, |
| SveScatterStoreSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, false); |
| } |
| case 0x3: |
| if (esizeIs32) { |
| break; |
| } else { |
| return new SveIndexedMemSV<uint64_t, uint64_t, |
| SveScatterStoreSVMicroop, |
| SveFirstFaultWritebackMicroop>( |
| mn, machInst, MemWriteOp, zt, pg, rn, zm, |
| offsetIs32, offsetIsSigned, offsetIsScaled, false); |
| } |
| } |
| return new Unknown64(machInst); |
| } |
| |
| }}; |
| |
| |
| let {{ |
| |
| header_output = '' |
| exec_output = '' |
| decoders = { 'Generic': {} } |
| |
| SPAlignmentCheckCode = ''' |
| if (this->baseIsSP && bits(XBase, 3, 0) && |
| SPAlignmentCheckEnabled(xc->tcBase())) { |
| return std::make_shared<SPAlignmentFault>(); |
| } |
| ''' |
| |
| def emitSveMemFillSpill(isPred): |
| global header_output, exec_output, decoders |
| eaCode = SPAlignmentCheckCode + ''' |
| int memAccessSize = %(memacc_size)s; |
| EA = XBase + ((int64_t) imm * %(memacc_size)s)''' % { |
| 'memacc_size': 'eCount / 8' if isPred else 'eCount'} |
| loadRdEnableCode = ''' |
| auto rdEn = std::vector<bool>(); |
| ''' |
| if isPred: |
| loadMemAccCode = ''' |
| int index = 0; |
| uint8_t byte; |
| for (int i = 0; i < eCount / 8; i++) { |
| byte = memDataView[i]; |
| for (int j = 0; j < 8; j++, index++) { |
| PDest_x[index] = (byte >> j) & 1; |
| } |
| } |
| ''' |
| storeMemAccCode = ''' |
| int index = 0; |
| uint8_t byte; |
| for (int i = 0; i < eCount / 8; i++) { |
| byte = 0; |
| for (int j = 0; j < 8; j++, index++) { |
| byte |= PDest_x[index] << j; |
| } |
| memDataView[i] = byte; |
| } |
| ''' |
| storeWrEnableCode = ''' |
| auto wrEn = std::vector<bool>(eCount / 8, true); |
| ''' |
| else: |
| loadMemAccCode = ''' |
| for (int i = 0; i < eCount; i++) { |
| AA64FpDest_x[i] = memDataView[i]; |
| } |
| ''' |
| storeMemAccCode = ''' |
| for (int i = 0; i < eCount; i++) { |
| memDataView[i] = AA64FpDest_x[i]; |
| } |
| ''' |
| storeWrEnableCode = ''' |
| auto wrEn = std::vector<bool>(sizeof(MemElemType) * eCount, true); |
| ''' |
| loadIop = InstObjParams('ldr', |
| 'SveLdrPred' if isPred else 'SveLdrVec', |
| 'SveMemPredFillSpill' if isPred else 'SveMemVecFillSpill', |
| {'tpl_header': '', |
| 'tpl_args': '', |
| 'memacc_code': loadMemAccCode, |
| 'ea_code' : sveEnabledCheckCode + eaCode, |
| 'rden_code' : loadRdEnableCode, |
| 'fault_code' : '', |
| 'fa_code' : ''}, |
| ['IsMemRef', 'IsLoad']) |
| storeIop = InstObjParams('str', |
| 'SveStrPred' if isPred else 'SveStrVec', |
| 'SveMemPredFillSpill' if isPred else 'SveMemVecFillSpill', |
| {'tpl_header': '', |
| 'tpl_args': '', |
| 'wren_code': storeWrEnableCode, |
| 'memacc_code': storeMemAccCode, |
| 'ea_code' : sveEnabledCheckCode + eaCode, |
| 'fa_code' : ''}, |
| ['IsMemRef', 'IsStore']) |
| header_output += SveMemFillSpillOpDeclare.subst(loadIop) |
| header_output += SveMemFillSpillOpDeclare.subst(storeIop) |
| exec_output += ( |
| SveContigLoadExecute.subst(loadIop) + |
| SveContigLoadInitiateAcc.subst(loadIop) + |
| SveContigLoadCompleteAcc.subst(loadIop) + |
| SveContigStoreExecute.subst(storeIop) + |
| SveContigStoreInitiateAcc.subst(storeIop) + |
| SveContigStoreCompleteAcc.subst(storeIop)) |
| |
| loadTplArgs = ( |
| ('uint8_t', 'uint8_t'), |
| ('uint16_t', 'uint8_t'), |
| ('uint32_t', 'uint8_t'), |
| ('uint64_t', 'uint8_t'), |
| ('int64_t', 'int32_t'), |
| ('uint16_t', 'uint16_t'), |
| ('uint32_t', 'uint16_t'), |
| ('uint64_t', 'uint16_t'), |
| ('int64_t', 'int16_t'), |
| ('int32_t', 'int16_t'), |
| ('uint32_t', 'uint32_t'), |
| ('uint64_t', 'uint32_t'), |
| ('int64_t', 'int8_t'), |
| ('int32_t', 'int8_t'), |
| ('int16_t', 'int8_t'), |
| ('uint64_t', 'uint64_t'), |
| ) |
| |
| storeTplArgs = ( |
| ('uint8_t', 'uint8_t'), |
| ('uint16_t', 'uint8_t'), |
| ('uint32_t', 'uint8_t'), |
| ('uint64_t', 'uint8_t'), |
| ('uint16_t', 'uint16_t'), |
| ('uint32_t', 'uint16_t'), |
| ('uint64_t', 'uint16_t'), |
| ('uint32_t', 'uint32_t'), |
| ('uint64_t', 'uint32_t'), |
| ('uint64_t', 'uint64_t'), |
| ) |
| |
| gatherLoadTplArgs = ( |
| ('int32_t', 'int8_t'), |
| ('int64_t', 'int8_t'), |
| ('uint32_t', 'uint8_t'), |
| ('uint64_t', 'uint8_t'), |
| ('int32_t', 'int16_t'), |
| ('int64_t', 'int16_t'), |
| ('uint32_t', 'uint16_t'), |
| ('uint64_t', 'uint16_t'), |
| ('int64_t', 'int32_t'), |
| ('uint32_t', 'uint32_t'), |
| ('uint64_t', 'uint32_t'), |
| ('uint64_t', 'uint64_t'), |
| ) |
| |
| scatterStoreTplArgs = ( |
| ('uint32_t', 'uint8_t'), |
| ('uint64_t', 'uint8_t'), |
| ('uint32_t', 'uint16_t'), |
| ('uint64_t', 'uint16_t'), |
| ('uint32_t', 'uint32_t'), |
| ('uint64_t', 'uint32_t'), |
| ('uint64_t', 'uint64_t'), |
| ) |
| |
| # Generates definitions for SVE contiguous loads |
| def emitSveContigMemInsts(offsetIsImm): |
| global header_output, exec_output, decoders |
| # First-faulting instructions only have a scalar plus scalar form, |
| # while non-faulting instructions only a scalar plus immediate form, so |
| # `offsetIsImm` is used to determine which class of instructions is |
| # generated |
| firstFaulting = not offsetIsImm |
| tplHeader = 'template <class RegElemType, class MemElemType>' |
| tplArgs = '<RegElemType, MemElemType>' |
| eaCode = SPAlignmentCheckCode + ''' |
| int memAccessSize = eCount * sizeof(MemElemType); |
| EA = XBase + ''' |
| if offsetIsImm: |
| eaCode += '((int64_t) this->imm * eCount * sizeof(MemElemType))' |
| else: |
| eaCode += '(XOffset * sizeof(MemElemType));' |
| loadRdEnableCode = ''' |
| auto rdEn = std::vector<bool>(sizeof(MemElemType) * eCount, true); |
| for (int i = 0; i < eCount; i++) { |
| if (!GpOp_x[i]) { |
| for (int j = 0; j < sizeof(MemElemType); j++) { |
| rdEn[sizeof(MemElemType) * i + j] = false; |
| } |
| } |
| } |
| ''' |
| loadMemAccCode = ''' |
| for (int i = 0; i < eCount; i++) { |
| if (GpOp_x[i]) { |
| AA64FpDest_x[i] = memDataView[i]; |
| } else { |
| AA64FpDest_x[i] = 0; |
| } |
| } |
| ''' |
| storeMemAccCode = ''' |
| for (int i = 0; i < eCount; i++) { |
| if (GpOp_x[i]) { |
| memDataView[i] = AA64FpDest_x[i]; |
| } else { |
| memDataView[i] = 0; |
| for (int j = 0; j < sizeof(MemElemType); j++) { |
| wrEn[sizeof(MemElemType) * i + j] = false; |
| } |
| } |
| } |
| ''' |
| storeWrEnableCode = ''' |
| auto wrEn = std::vector<bool>(sizeof(MemElemType) * eCount, true); |
| ''' |
| ffrReadBackCode = ''' |
| auto& firstFaultReg = Ffr;''' |
| fautlingLoadmemAccCode = ''' |
| for (int i = 0; i < eCount; i++) { |
| if (GpOp_x[i] && firstFaultReg[i * sizeof(RegElemType)]) { |
| AA64FpDest_x[i] = memDataView[i]; |
| } else { |
| AA64FpDest_x[i] = 0; |
| } |
| } |
| ''' |
| nonFaultingCode = 'true ||' |
| faultCode = ''' |
| Addr fault_addr; |
| if (fault == NoFault || getFaultVAddr(fault, fault_addr)) { |
| unsigned fault_elem_index; |
| if (fault != NoFault) { |
| assert(fault_addr >= EA); |
| fault_elem_index = (fault_addr - EA) / sizeof(MemElemType); |
| } else { |
| fault_elem_index = eCount + 1; |
| } |
| int first_active_index; |
| for (first_active_index = 0; |
| first_active_index < eCount && !(GpOp_x[first_active_index]); |
| first_active_index++); |
| if (%s first_active_index < fault_elem_index) { |
| for (int i = 0; i < eCount; i++) { |
| for (int j = 0; j < sizeof(RegElemType); j++) { |
| if (i < fault_elem_index) { |
| Ffr_ub[i * sizeof(RegElemType) + j] = FfrAux_x[i]; |
| } else { |
| Ffr_ub[i * sizeof(RegElemType) + j] = 0; |
| } |
| } |
| } |
| fault = NoFault; |
| if (first_active_index >= fault_elem_index) { |
| // non-faulting load needs this |
| xc->setMemAccPredicate(false); |
| } |
| } |
| } |
| ''' % ('' if firstFaulting else nonFaultingCode) |
| |
| loadIop = InstObjParams('ld1', |
| 'SveContigLoadSI' if offsetIsImm else 'SveContigLoadSS', |
| 'SveContigMemSI' if offsetIsImm else 'SveContigMemSS', |
| {'tpl_header': tplHeader, |
| 'tpl_args': tplArgs, |
| 'rden_code' : loadRdEnableCode, |
| 'memacc_code': loadMemAccCode, |
| 'ea_code' : sveEnabledCheckCode + eaCode, |
| 'fault_code' : '', |
| 'fa_code' : ''}, |
| ['IsMemRef', 'IsLoad']) |
| storeIop = InstObjParams('st1', |
| 'SveContigStoreSI' if offsetIsImm else 'SveContigStoreSS', |
| 'SveContigMemSI' if offsetIsImm else 'SveContigMemSS', |
| {'tpl_header': tplHeader, |
| 'tpl_args': tplArgs, |
| 'wren_code': storeWrEnableCode, |
| 'memacc_code': storeMemAccCode, |
| 'ea_code' : sveEnabledCheckCode + eaCode, |
| 'fa_code' : ''}, |
| ['IsMemRef', 'IsStore']) |
| faultIop = InstObjParams('ldff1' if firstFaulting else 'ldnf1', |
| 'SveContigFFLoadSS' if firstFaulting else 'SveContigNFLoadSI', |
| 'SveContigMemSS' if firstFaulting else 'SveContigMemSI', |
| {'tpl_header': tplHeader, |
| 'tpl_args': tplArgs, |
| 'rden_code' : loadRdEnableCode, |
| 'memacc_code': fautlingLoadmemAccCode, |
| 'ea_code' : sveEnabledCheckCode + eaCode, |
| 'fault_code' : faultCode, |
| 'fa_code' : ''}, |
| ['IsMemRef', 'IsLoad']) |
| faultIop.snippets['memacc_code'] = (ffrReadBackCode + |
| faultIop.snippets['memacc_code']) |
| if offsetIsImm: |
| header_output += SveContigMemSIOpDeclare.subst(loadIop) |
| header_output += SveContigMemSIOpDeclare.subst(storeIop) |
| header_output += SveContigMemSIOpDeclare.subst(faultIop) |
| else: |
| header_output += SveContigMemSSOpDeclare.subst(loadIop) |
| header_output += SveContigMemSSOpDeclare.subst(storeIop) |
| header_output += SveContigMemSSOpDeclare.subst(faultIop) |
| exec_output += ( |
| SveContigLoadExecute.subst(loadIop) + |
| SveContigLoadInitiateAcc.subst(loadIop) + |
| SveContigLoadCompleteAcc.subst(loadIop) + |
| SveContigStoreExecute.subst(storeIop) + |
| SveContigStoreInitiateAcc.subst(storeIop) + |
| SveContigStoreCompleteAcc.subst(storeIop) + |
| SveContigLoadExecute.subst(faultIop) + |
| SveContigLoadInitiateAcc.subst(faultIop) + |
| SveContigLoadCompleteAcc.subst(faultIop)) |
| |
| for args in loadTplArgs: |
| substDict = {'tpl_args': '<%s>' % ', '.join(args), |
| 'class_name': 'SveContigLoadSI' if offsetIsImm |
| else 'SveContigLoadSS'} |
| exec_output += SveContigMemExecDeclare.subst(substDict) |
| for args in storeTplArgs: |
| substDict = {'tpl_args': '<%s>' % ', '.join(args), |
| 'class_name': 'SveContigStoreSI' if offsetIsImm |
| else 'SveContigStoreSS'} |
| exec_output += SveContigMemExecDeclare.subst(substDict) |
| for args in loadTplArgs: |
| substDict = {'tpl_args': '<%s>' % ', '.join(args), |
| 'class_name': 'SveContigFFLoadSS' if firstFaulting |
| else 'SveContigNFLoadSI'} |
| exec_output += SveContigMemExecDeclare.subst(substDict) |
| |
| |
| # Generates definitions for SVE load-and-replicate instructions |
| def emitSveLoadAndRepl(): |
| global header_output, exec_output, decoders |
| tplHeader = 'template <class RegElemType, class MemElemType>' |
| tplArgs = '<RegElemType, MemElemType>' |
| eaCode = SPAlignmentCheckCode + ''' |
| EA = XBase + imm * sizeof(MemElemType);''' |
| memAccCode = ''' |
| for (int i = 0; i < eCount; i++) { |
| if (GpOp_x[i]) { |
| AA64FpDest_x[i] = memData; |
| } else { |
| AA64FpDest_x[i] = 0; |
| } |
| } |
| ''' |
| iop = InstObjParams('ld1r', |
| 'SveLoadAndRepl', |
| 'SveContigMemSI', |
| {'tpl_header': tplHeader, |
| 'tpl_args': tplArgs, |
| 'memacc_code': memAccCode, |
| 'ea_code' : sveEnabledCheckCode + eaCode, |
| 'fa_code' : ''}, |
| ['IsMemRef', 'IsLoad']) |
| header_output += SveContigMemSIOpDeclare.subst(iop) |
| exec_output += ( |
| SveLoadAndReplExecute.subst(iop) + |
| SveLoadAndReplInitiateAcc.subst(iop) + |
| SveLoadAndReplCompleteAcc.subst(iop)) |
| for args in loadTplArgs: |
| substDict = {'tpl_args': '<%s>' % ', '.join(args), |
| 'class_name': 'SveLoadAndRepl'} |
| exec_output += SveContigMemExecDeclare.subst(substDict) |
| |
| class IndexedAddrForm: |
| VEC_PLUS_IMM = 0 |
| SCA_PLUS_VEC = 1 |
| |
| # Generates definitions for the transfer microops of SVE indexed memory |
| # operations (gather loads, scatter stores) |
| def emitSveIndexedMemMicroops(indexed_addr_form): |
| assert indexed_addr_form in (IndexedAddrForm.VEC_PLUS_IMM, |
| IndexedAddrForm.SCA_PLUS_VEC) |
| global header_output, exec_output, decoders |
| tplHeader = 'template <class RegElemType, class MemElemType>' |
| tplArgs = '<RegElemType, MemElemType>' |
| if indexed_addr_form == IndexedAddrForm.VEC_PLUS_IMM: |
| eaCode = ''' |
| EA = AA64FpBase_x[elemIndex] + imm * sizeof(MemElemType)''' |
| else: |
| eaCode = ''' |
| uint64_t offset = AA64FpOffset_x[elemIndex]; |
| if (offsetIs32) { |
| offset &= (1ULL << 32) - 1; |
| } |
| if (offsetIsSigned) { |
| offset = sext<32>(offset); |
| } |
| if (offsetIsScaled) { |
| offset *= sizeof(MemElemType); |
| } |
| EA = XBase + offset''' |
| loadMemAccCode = ''' |
| AA64FpDest_x[elemIndex] = memData; |
| ''' |
| storeMemAccCode = ''' |
| memData = AA64FpDest_x[elemIndex]; |
| ''' |
| predCheckCode = 'GpOp_x[index]' |
| faultStatusSetCode = 'PUreg0_x[elemIndex] = 1;' |
| faultStatusResetCode = 'PUreg0_x[elemIndex] = 0;' |
| loadIop = InstObjParams('ld1', |
| ('SveGatherLoadVIMicroop' |
| if indexed_addr_form == IndexedAddrForm.VEC_PLUS_IMM |
| else 'SveGatherLoadSVMicroop'), |
| 'MicroOp', |
| {'tpl_header': tplHeader, |
| 'tpl_args': tplArgs, |
| 'memacc_code': loadMemAccCode, |
| 'ea_code' : sveEnabledCheckCode + eaCode, |
| 'fault_status_set_code' : faultStatusSetCode, |
| 'fault_status_reset_code' : faultStatusResetCode, |
| 'pred_check_code' : predCheckCode, |
| 'fa_code' : ''}, |
| ['IsMicroop', 'IsMemRef', 'IsLoad']) |
| storeIop = InstObjParams('st1', |
| ('SveScatterStoreVIMicroop' |
| if indexed_addr_form == IndexedAddrForm.VEC_PLUS_IMM |
| else 'SveScatterStoreSVMicroop'), |
| 'MicroOp', |
| {'tpl_header': tplHeader, |
| 'tpl_args': tplArgs, |
| 'memacc_code': storeMemAccCode, |
| 'ea_code' : sveEnabledCheckCode + eaCode, |
| 'pred_check_code' : predCheckCode, |
| 'fa_code' : ''}, |
| ['IsMicroop', 'IsMemRef', 'IsStore']) |
| if indexed_addr_form == IndexedAddrForm.VEC_PLUS_IMM: |
| header_output += SveIndexedMemVIMicroopDeclare.subst(loadIop) |
| header_output += SveIndexedMemVIMicroopDeclare.subst(storeIop) |
| else: |
| header_output += SveIndexedMemSVMicroopDeclare.subst(loadIop) |
| header_output += SveIndexedMemSVMicroopDeclare.subst(storeIop) |
| exec_output += ( |
| SveGatherLoadMicroopExecute.subst(loadIop) + |
| SveGatherLoadMicroopInitiateAcc.subst(loadIop) + |
| SveGatherLoadMicroopCompleteAcc.subst(loadIop) + |
| SveScatterStoreMicroopExecute.subst(storeIop) + |
| SveScatterStoreMicroopInitiateAcc.subst(storeIop) + |
| SveScatterStoreMicroopCompleteAcc.subst(storeIop)) |
| for args in gatherLoadTplArgs: |
| substDict = {'tpl_args': '<%s>' % ', '.join(args), |
| 'class_name': ( |
| 'SveGatherLoadVIMicroop' |
| if indexed_addr_form == \ |
| IndexedAddrForm.VEC_PLUS_IMM |
| else 'SveGatherLoadSVMicroop')} |
| # TODO: this should become SveMemExecDeclare |
| exec_output += SveContigMemExecDeclare.subst(substDict) |
| for args in scatterStoreTplArgs: |
| substDict = {'tpl_args': '<%s>' % ', '.join(args), |
| 'class_name': ( |
| 'SveScatterStoreVIMicroop' |
| if indexed_addr_form == \ |
| IndexedAddrForm.VEC_PLUS_IMM |
| else 'SveScatterStoreSVMicroop')} |
| # TODO: this should become SveMemExecDeclare |
| exec_output += SveContigMemExecDeclare.subst(substDict) |
| |
| firstFaultTplArgs = ('int32_t', 'int64_t', 'uint32_t', 'uint64_t') |
| |
| def emitSveFirstFaultWritebackMicroop(): |
| global header_output, exec_output, decoders |
| tplHeader = 'template <class RegElemType>' |
| tplArgs = '<RegElemType>' |
| faultStatusCheckCode = 'PUreg0_x[index]' |
| firstFaultResetCode = ''' |
| for(int j = 0; j < sizeof(RegElemType); j++) { |
| Ffr_ub[index * sizeof(RegElemType) + j] = 0; |
| } |
| ''' |
| firstFaultForwardCode = ''' |
| for(int j = 0; j < sizeof(RegElemType); j++) { |
| Ffr_ub[index * sizeof(RegElemType) + j] = FfrAux_x[index]; |
| } |
| ''' |
| iop = InstObjParams('ldff1', |
| 'SveFirstFaultWritebackMicroop', |
| 'MicroOp', |
| {'tpl_header': tplHeader, |
| 'tpl_args': tplArgs, |
| 'fault_status_check_code' : faultStatusCheckCode, |
| 'first_fault_reset_code' : firstFaultResetCode, |
| 'first_fault_forward_code' : firstFaultForwardCode}, |
| ['IsMicroop']) |
| header_output += SveFirstFaultWritebackMicroopDeclare.subst(iop) |
| exec_output += SveFirstFaultWritebackMicroopExecute.subst(iop) |
| for args in firstFaultTplArgs: |
| substDict = {'targs': args, |
| 'class_name' : 'SveFirstFaultWritebackMicroop' } |
| exec_output += SveOpExecDeclare.subst(substDict) |
| |
| # Generates definitions for the first microop of SVE gather loads, required |
| # to propagate the source vector register to the transfer microops |
| def emitSveGatherLoadCpySrcVecMicroop(): |
| global header_output, exec_output, decoders |
| code = sveEnabledCheckCode + ''' |
| unsigned eCount = ArmStaticInst::getCurSveVecLen<uint8_t>( |
| xc->tcBase()); |
| for (unsigned i = 0; i < eCount; i++) { |
| AA64FpUreg0_ub[i] = AA64FpOp1_ub[i]; |
| }''' |
| iop = InstObjParams('ld1', |
| 'SveGatherLoadCpySrcVecMicroop', |
| 'MicroOp', |
| {'code': code}, |
| ['IsMicroop']) |
| header_output += SveGatherLoadCpySrcVecMicroopDeclare.subst(iop) |
| exec_output += SveGatherLoadCpySrcVecMicroopExecute.subst(iop) |
| |
| # LD1[S]{B,H,W,D} (scalar plus immediate) |
| # ST1[S]{B,H,W,D} (scalar plus immediate) |
| # LDNF1[S]{B,H,W,D} (scalar plus immediate) |
| emitSveContigMemInsts(True) |
| # LD1[S]{B,H,W,D} (scalar plus scalar) |
| # ST1[S]{B,H,W,D} (scalar plus scalar) |
| # LDFF1[S]{B,H,W,D} (scalar plus vector) |
| emitSveContigMemInsts(False) |
| |
| # LD1R[S]{B,H,W,D} |
| emitSveLoadAndRepl() |
| |
| # LDR (predicate), STR (predicate) |
| emitSveMemFillSpill(True) |
| # LDR (vector), STR (vector) |
| emitSveMemFillSpill(False) |
| |
| # LD1[S]{B,H,W,D} (vector plus immediate) |
| # ST1[S]{B,H,W,D} (vector plus immediate) |
| # LDFF1[S]{B,H,W,D} (scalar plus immediate) |
| emitSveIndexedMemMicroops(IndexedAddrForm.VEC_PLUS_IMM) |
| # LD1[S]{B,H,W,D} (scalar plus vector) |
| # ST1[S]{B,H,W,D} (scalar plus vector) |
| # LDFF1[S]{B,H,W,D} (scalar plus vector) |
| emitSveIndexedMemMicroops(IndexedAddrForm.SCA_PLUS_VEC) |
| |
| # FFR writeback microop for gather loads |
| emitSveFirstFaultWritebackMicroop() |
| |
| # Source vector copy microop for gather loads |
| emitSveGatherLoadCpySrcVecMicroop() |
| }}; |