xref: /freebsd-src/contrib/llvm-project/llvm/lib/Target/X86/X86CompressEVEX.cpp (revision 7a6dacaca14b62ca4b74406814becb87a3fefac0)
11db9f3b2SDimitry Andric //===- X86CompressEVEX.cpp ------------------------------------------------===//
21db9f3b2SDimitry Andric //
31db9f3b2SDimitry Andric // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
41db9f3b2SDimitry Andric // See https://llvm.org/LICENSE.txt for license information.
51db9f3b2SDimitry Andric // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
61db9f3b2SDimitry Andric //
71db9f3b2SDimitry Andric //===----------------------------------------------------------------------===//
81db9f3b2SDimitry Andric //
91db9f3b2SDimitry Andric // This pass compresses instructions from EVEX space to legacy/VEX/EVEX space
101db9f3b2SDimitry Andric // when possible in order to reduce code size or facilitate HW decoding.
111db9f3b2SDimitry Andric //
121db9f3b2SDimitry Andric // Possible compression:
131db9f3b2SDimitry Andric //   a. AVX512 instruction (EVEX) -> AVX instruction (VEX)
141db9f3b2SDimitry Andric //   b. Promoted instruction (EVEX) -> pre-promotion instruction (legacy/VEX)
151db9f3b2SDimitry Andric //   c. NDD (EVEX) -> non-NDD (legacy)
161db9f3b2SDimitry Andric //   d. NF_ND (EVEX) -> NF (EVEX)
171db9f3b2SDimitry Andric //
181db9f3b2SDimitry Andric // Compression a, b and c can always reduce code size, with some exceptions
191db9f3b2SDimitry Andric // such as promoted 16-bit CRC32 which is as long as the legacy version.
201db9f3b2SDimitry Andric //
211db9f3b2SDimitry Andric // legacy:
221db9f3b2SDimitry Andric //   crc32w %si, %eax ## encoding: [0x66,0xf2,0x0f,0x38,0xf1,0xc6]
231db9f3b2SDimitry Andric // promoted:
241db9f3b2SDimitry Andric //   crc32w %si, %eax ## encoding: [0x62,0xf4,0x7d,0x08,0xf1,0xc6]
251db9f3b2SDimitry Andric //
261db9f3b2SDimitry Andric // From performance perspective, these should be same (same uops and same EXE
271db9f3b2SDimitry Andric // ports). From a FMV perspective, an older legacy encoding is preferred b/c it
281db9f3b2SDimitry Andric // can execute in more places (broader HW install base). So we will still do
291db9f3b2SDimitry Andric // the compression.
301db9f3b2SDimitry Andric //
311db9f3b2SDimitry Andric // Compression d can help hardware decode (HW may skip reading the NDD
321db9f3b2SDimitry Andric // register) although the instruction length remains unchanged.
331db9f3b2SDimitry Andric //===----------------------------------------------------------------------===//
341db9f3b2SDimitry Andric 
351db9f3b2SDimitry Andric #include "MCTargetDesc/X86BaseInfo.h"
361db9f3b2SDimitry Andric #include "MCTargetDesc/X86InstComments.h"
371db9f3b2SDimitry Andric #include "X86.h"
381db9f3b2SDimitry Andric #include "X86InstrInfo.h"
391db9f3b2SDimitry Andric #include "X86Subtarget.h"
401db9f3b2SDimitry Andric #include "llvm/ADT/StringRef.h"
411db9f3b2SDimitry Andric #include "llvm/CodeGen/MachineFunction.h"
421db9f3b2SDimitry Andric #include "llvm/CodeGen/MachineFunctionPass.h"
431db9f3b2SDimitry Andric #include "llvm/CodeGen/MachineInstr.h"
441db9f3b2SDimitry Andric #include "llvm/CodeGen/MachineOperand.h"
451db9f3b2SDimitry Andric #include "llvm/MC/MCInstrDesc.h"
461db9f3b2SDimitry Andric #include "llvm/Pass.h"
471db9f3b2SDimitry Andric #include <atomic>
481db9f3b2SDimitry Andric #include <cassert>
491db9f3b2SDimitry Andric #include <cstdint>
501db9f3b2SDimitry Andric 
511db9f3b2SDimitry Andric using namespace llvm;
521db9f3b2SDimitry Andric 
531db9f3b2SDimitry Andric // Including the generated EVEX compression tables.
541db9f3b2SDimitry Andric struct X86CompressEVEXTableEntry {
551db9f3b2SDimitry Andric   uint16_t OldOpc;
561db9f3b2SDimitry Andric   uint16_t NewOpc;
571db9f3b2SDimitry Andric 
581db9f3b2SDimitry Andric   bool operator<(const X86CompressEVEXTableEntry &RHS) const {
591db9f3b2SDimitry Andric     return OldOpc < RHS.OldOpc;
601db9f3b2SDimitry Andric   }
611db9f3b2SDimitry Andric 
621db9f3b2SDimitry Andric   friend bool operator<(const X86CompressEVEXTableEntry &TE, unsigned Opc) {
631db9f3b2SDimitry Andric     return TE.OldOpc < Opc;
641db9f3b2SDimitry Andric   }
651db9f3b2SDimitry Andric };
661db9f3b2SDimitry Andric #include "X86GenCompressEVEXTables.inc"
671db9f3b2SDimitry Andric 
681db9f3b2SDimitry Andric #define COMP_EVEX_DESC "Compressing EVEX instrs when possible"
691db9f3b2SDimitry Andric #define COMP_EVEX_NAME "x86-compress-evex"
701db9f3b2SDimitry Andric 
711db9f3b2SDimitry Andric #define DEBUG_TYPE COMP_EVEX_NAME
721db9f3b2SDimitry Andric 
731db9f3b2SDimitry Andric namespace {
741db9f3b2SDimitry Andric 
751db9f3b2SDimitry Andric class CompressEVEXPass : public MachineFunctionPass {
761db9f3b2SDimitry Andric public:
771db9f3b2SDimitry Andric   static char ID;
781db9f3b2SDimitry Andric   CompressEVEXPass() : MachineFunctionPass(ID) {}
791db9f3b2SDimitry Andric   StringRef getPassName() const override { return COMP_EVEX_DESC; }
801db9f3b2SDimitry Andric 
811db9f3b2SDimitry Andric   bool runOnMachineFunction(MachineFunction &MF) override;
821db9f3b2SDimitry Andric 
831db9f3b2SDimitry Andric   // This pass runs after regalloc and doesn't support VReg operands.
841db9f3b2SDimitry Andric   MachineFunctionProperties getRequiredProperties() const override {
851db9f3b2SDimitry Andric     return MachineFunctionProperties().set(
861db9f3b2SDimitry Andric         MachineFunctionProperties::Property::NoVRegs);
871db9f3b2SDimitry Andric   }
881db9f3b2SDimitry Andric };
891db9f3b2SDimitry Andric 
901db9f3b2SDimitry Andric } // end anonymous namespace
911db9f3b2SDimitry Andric 
921db9f3b2SDimitry Andric char CompressEVEXPass::ID = 0;
931db9f3b2SDimitry Andric 
941db9f3b2SDimitry Andric static bool usesExtendedRegister(const MachineInstr &MI) {
951db9f3b2SDimitry Andric   auto isHiRegIdx = [](unsigned Reg) {
961db9f3b2SDimitry Andric     // Check for XMM register with indexes between 16 - 31.
971db9f3b2SDimitry Andric     if (Reg >= X86::XMM16 && Reg <= X86::XMM31)
981db9f3b2SDimitry Andric       return true;
991db9f3b2SDimitry Andric     // Check for YMM register with indexes between 16 - 31.
1001db9f3b2SDimitry Andric     if (Reg >= X86::YMM16 && Reg <= X86::YMM31)
1011db9f3b2SDimitry Andric       return true;
1021db9f3b2SDimitry Andric     // Check for GPR with indexes between 16 - 31.
1031db9f3b2SDimitry Andric     if (X86II::isApxExtendedReg(Reg))
1041db9f3b2SDimitry Andric       return true;
1051db9f3b2SDimitry Andric     return false;
1061db9f3b2SDimitry Andric   };
1071db9f3b2SDimitry Andric 
1081db9f3b2SDimitry Andric   // Check that operands are not ZMM regs or
1091db9f3b2SDimitry Andric   // XMM/YMM regs with hi indexes between 16 - 31.
1101db9f3b2SDimitry Andric   for (const MachineOperand &MO : MI.explicit_operands()) {
1111db9f3b2SDimitry Andric     if (!MO.isReg())
1121db9f3b2SDimitry Andric       continue;
1131db9f3b2SDimitry Andric 
1141db9f3b2SDimitry Andric     Register Reg = MO.getReg();
1151db9f3b2SDimitry Andric     assert(!X86II::isZMMReg(Reg) &&
1161db9f3b2SDimitry Andric            "ZMM instructions should not be in the EVEX->VEX tables");
1171db9f3b2SDimitry Andric     if (isHiRegIdx(Reg))
1181db9f3b2SDimitry Andric       return true;
1191db9f3b2SDimitry Andric   }
1201db9f3b2SDimitry Andric 
1211db9f3b2SDimitry Andric   return false;
1221db9f3b2SDimitry Andric }
1231db9f3b2SDimitry Andric 
1241db9f3b2SDimitry Andric // Do any custom cleanup needed to finalize the conversion.
1251db9f3b2SDimitry Andric static bool performCustomAdjustments(MachineInstr &MI, unsigned NewOpc) {
1261db9f3b2SDimitry Andric   (void)NewOpc;
1271db9f3b2SDimitry Andric   unsigned Opc = MI.getOpcode();
1281db9f3b2SDimitry Andric   switch (Opc) {
1291db9f3b2SDimitry Andric   case X86::VALIGNDZ128rri:
1301db9f3b2SDimitry Andric   case X86::VALIGNDZ128rmi:
1311db9f3b2SDimitry Andric   case X86::VALIGNQZ128rri:
1321db9f3b2SDimitry Andric   case X86::VALIGNQZ128rmi: {
1331db9f3b2SDimitry Andric     assert((NewOpc == X86::VPALIGNRrri || NewOpc == X86::VPALIGNRrmi) &&
1341db9f3b2SDimitry Andric            "Unexpected new opcode!");
1351db9f3b2SDimitry Andric     unsigned Scale =
1361db9f3b2SDimitry Andric         (Opc == X86::VALIGNQZ128rri || Opc == X86::VALIGNQZ128rmi) ? 8 : 4;
1371db9f3b2SDimitry Andric     MachineOperand &Imm = MI.getOperand(MI.getNumExplicitOperands() - 1);
1381db9f3b2SDimitry Andric     Imm.setImm(Imm.getImm() * Scale);
1391db9f3b2SDimitry Andric     break;
1401db9f3b2SDimitry Andric   }
1411db9f3b2SDimitry Andric   case X86::VSHUFF32X4Z256rmi:
1421db9f3b2SDimitry Andric   case X86::VSHUFF32X4Z256rri:
1431db9f3b2SDimitry Andric   case X86::VSHUFF64X2Z256rmi:
1441db9f3b2SDimitry Andric   case X86::VSHUFF64X2Z256rri:
1451db9f3b2SDimitry Andric   case X86::VSHUFI32X4Z256rmi:
1461db9f3b2SDimitry Andric   case X86::VSHUFI32X4Z256rri:
1471db9f3b2SDimitry Andric   case X86::VSHUFI64X2Z256rmi:
1481db9f3b2SDimitry Andric   case X86::VSHUFI64X2Z256rri: {
1491db9f3b2SDimitry Andric     assert((NewOpc == X86::VPERM2F128rr || NewOpc == X86::VPERM2I128rr ||
1501db9f3b2SDimitry Andric             NewOpc == X86::VPERM2F128rm || NewOpc == X86::VPERM2I128rm) &&
1511db9f3b2SDimitry Andric            "Unexpected new opcode!");
1521db9f3b2SDimitry Andric     MachineOperand &Imm = MI.getOperand(MI.getNumExplicitOperands() - 1);
1531db9f3b2SDimitry Andric     int64_t ImmVal = Imm.getImm();
1541db9f3b2SDimitry Andric     // Set bit 5, move bit 1 to bit 4, copy bit 0.
1551db9f3b2SDimitry Andric     Imm.setImm(0x20 | ((ImmVal & 2) << 3) | (ImmVal & 1));
1561db9f3b2SDimitry Andric     break;
1571db9f3b2SDimitry Andric   }
1581db9f3b2SDimitry Andric   case X86::VRNDSCALEPDZ128rri:
1591db9f3b2SDimitry Andric   case X86::VRNDSCALEPDZ128rmi:
1601db9f3b2SDimitry Andric   case X86::VRNDSCALEPSZ128rri:
1611db9f3b2SDimitry Andric   case X86::VRNDSCALEPSZ128rmi:
1621db9f3b2SDimitry Andric   case X86::VRNDSCALEPDZ256rri:
1631db9f3b2SDimitry Andric   case X86::VRNDSCALEPDZ256rmi:
1641db9f3b2SDimitry Andric   case X86::VRNDSCALEPSZ256rri:
1651db9f3b2SDimitry Andric   case X86::VRNDSCALEPSZ256rmi:
1661db9f3b2SDimitry Andric   case X86::VRNDSCALESDZr:
1671db9f3b2SDimitry Andric   case X86::VRNDSCALESDZm:
1681db9f3b2SDimitry Andric   case X86::VRNDSCALESSZr:
1691db9f3b2SDimitry Andric   case X86::VRNDSCALESSZm:
1701db9f3b2SDimitry Andric   case X86::VRNDSCALESDZr_Int:
1711db9f3b2SDimitry Andric   case X86::VRNDSCALESDZm_Int:
1721db9f3b2SDimitry Andric   case X86::VRNDSCALESSZr_Int:
1731db9f3b2SDimitry Andric   case X86::VRNDSCALESSZm_Int:
1741db9f3b2SDimitry Andric     const MachineOperand &Imm = MI.getOperand(MI.getNumExplicitOperands() - 1);
1751db9f3b2SDimitry Andric     int64_t ImmVal = Imm.getImm();
1761db9f3b2SDimitry Andric     // Ensure that only bits 3:0 of the immediate are used.
1771db9f3b2SDimitry Andric     if ((ImmVal & 0xf) != ImmVal)
1781db9f3b2SDimitry Andric       return false;
1791db9f3b2SDimitry Andric     break;
1801db9f3b2SDimitry Andric   }
1811db9f3b2SDimitry Andric 
1821db9f3b2SDimitry Andric   return true;
1831db9f3b2SDimitry Andric }
1841db9f3b2SDimitry Andric 
185*7a6dacacSDimitry Andric static bool isRedundantNewDataDest(MachineInstr &MI, const X86Subtarget &ST) {
186*7a6dacacSDimitry Andric   // $rbx = ADD64rr_ND $rbx, $rax / $rbx = ADD64rr_ND $rax, $rbx
187*7a6dacacSDimitry Andric   //   ->
188*7a6dacacSDimitry Andric   // $rbx = ADD64rr $rbx, $rax
189*7a6dacacSDimitry Andric   const MCInstrDesc &Desc = MI.getDesc();
190*7a6dacacSDimitry Andric   Register Reg0 = MI.getOperand(0).getReg();
191*7a6dacacSDimitry Andric   const MachineOperand &Op1 = MI.getOperand(1);
192*7a6dacacSDimitry Andric   if (!Op1.isReg())
193*7a6dacacSDimitry Andric     return false;
194*7a6dacacSDimitry Andric   Register Reg1 = Op1.getReg();
195*7a6dacacSDimitry Andric   if (Reg1 == Reg0)
196*7a6dacacSDimitry Andric     return true;
197*7a6dacacSDimitry Andric 
198*7a6dacacSDimitry Andric   // Op1 and Op2 may be commutable for ND instructions.
199*7a6dacacSDimitry Andric   if (!Desc.isCommutable() || Desc.getNumOperands() < 3 ||
200*7a6dacacSDimitry Andric       !MI.getOperand(2).isReg() || MI.getOperand(2).getReg() != Reg0)
201*7a6dacacSDimitry Andric     return false;
202*7a6dacacSDimitry Andric   // Opcode may change after commute, e.g. SHRD -> SHLD
203*7a6dacacSDimitry Andric   ST.getInstrInfo()->commuteInstruction(MI, false, 1, 2);
204*7a6dacacSDimitry Andric   return true;
205*7a6dacacSDimitry Andric }
206*7a6dacacSDimitry Andric 
2071db9f3b2SDimitry Andric static bool CompressEVEXImpl(MachineInstr &MI, const X86Subtarget &ST) {
2081db9f3b2SDimitry Andric   uint64_t TSFlags = MI.getDesc().TSFlags;
2091db9f3b2SDimitry Andric 
2101db9f3b2SDimitry Andric   // Check for EVEX instructions only.
2111db9f3b2SDimitry Andric   if ((TSFlags & X86II::EncodingMask) != X86II::EVEX)
2121db9f3b2SDimitry Andric     return false;
2131db9f3b2SDimitry Andric 
2141db9f3b2SDimitry Andric   // Instructions with mask or 512-bit vector can't be converted to VEX.
2151db9f3b2SDimitry Andric   if (TSFlags & (X86II::EVEX_K | X86II::EVEX_L2))
2161db9f3b2SDimitry Andric     return false;
2171db9f3b2SDimitry Andric 
2181db9f3b2SDimitry Andric   // EVEX_B has several meanings.
2191db9f3b2SDimitry Andric   // AVX512:
2201db9f3b2SDimitry Andric   //  register form: rounding control or SAE
2211db9f3b2SDimitry Andric   //  memory form: broadcast
2221db9f3b2SDimitry Andric   //
2231db9f3b2SDimitry Andric   // APX:
2241db9f3b2SDimitry Andric   //  MAP4: NDD
2251db9f3b2SDimitry Andric   //
2261db9f3b2SDimitry Andric   // For AVX512 cases, EVEX prefix is needed in order to carry this information
2271db9f3b2SDimitry Andric   // thus preventing the transformation to VEX encoding.
228*7a6dacacSDimitry Andric   bool IsND = X86II::hasNewDataDest(TSFlags);
2291db9f3b2SDimitry Andric   if (TSFlags & X86II::EVEX_B)
230*7a6dacacSDimitry Andric     if (!IsND || !isRedundantNewDataDest(MI, ST))
2311db9f3b2SDimitry Andric       return false;
2321db9f3b2SDimitry Andric 
2331db9f3b2SDimitry Andric   ArrayRef<X86CompressEVEXTableEntry> Table = ArrayRef(X86CompressEVEXTable);
2341db9f3b2SDimitry Andric 
2351db9f3b2SDimitry Andric   unsigned Opc = MI.getOpcode();
2361db9f3b2SDimitry Andric   const auto *I = llvm::lower_bound(Table, Opc);
237*7a6dacacSDimitry Andric   if (I == Table.end() || I->OldOpc != Opc) {
238*7a6dacacSDimitry Andric     assert(!IsND && "Missing entry for ND instruction");
2391db9f3b2SDimitry Andric     return false;
240*7a6dacacSDimitry Andric   }
2411db9f3b2SDimitry Andric 
242*7a6dacacSDimitry Andric   if (!IsND) {
243*7a6dacacSDimitry Andric     if (usesExtendedRegister(MI) || !checkPredicate(I->NewOpc, &ST) ||
2441db9f3b2SDimitry Andric         !performCustomAdjustments(MI, I->NewOpc))
2451db9f3b2SDimitry Andric       return false;
246*7a6dacacSDimitry Andric   }
2471db9f3b2SDimitry Andric 
2481db9f3b2SDimitry Andric   const MCInstrDesc &NewDesc = ST.getInstrInfo()->get(I->NewOpc);
2491db9f3b2SDimitry Andric   MI.setDesc(NewDesc);
250*7a6dacacSDimitry Andric   unsigned AsmComment;
251*7a6dacacSDimitry Andric   switch (NewDesc.TSFlags & X86II::EncodingMask) {
252*7a6dacacSDimitry Andric   case X86II::LEGACY:
253*7a6dacacSDimitry Andric     AsmComment = X86::AC_EVEX_2_LEGACY;
254*7a6dacacSDimitry Andric     break;
255*7a6dacacSDimitry Andric   case X86II::VEX:
256*7a6dacacSDimitry Andric     AsmComment = X86::AC_EVEX_2_VEX;
257*7a6dacacSDimitry Andric     break;
258*7a6dacacSDimitry Andric   case X86II::EVEX:
259*7a6dacacSDimitry Andric     AsmComment = X86::AC_EVEX_2_EVEX;
260*7a6dacacSDimitry Andric     assert(IsND && (NewDesc.TSFlags & X86II::EVEX_NF) &&
261*7a6dacacSDimitry Andric            "Unknown EVEX2EVEX compression");
262*7a6dacacSDimitry Andric     break;
263*7a6dacacSDimitry Andric   default:
264*7a6dacacSDimitry Andric     llvm_unreachable("Unknown EVEX compression");
265*7a6dacacSDimitry Andric   }
2661db9f3b2SDimitry Andric   MI.setAsmPrinterFlag(AsmComment);
267*7a6dacacSDimitry Andric   if (IsND)
268*7a6dacacSDimitry Andric     MI.tieOperands(0, 1);
269*7a6dacacSDimitry Andric 
2701db9f3b2SDimitry Andric   return true;
2711db9f3b2SDimitry Andric }
2721db9f3b2SDimitry Andric 
2731db9f3b2SDimitry Andric bool CompressEVEXPass::runOnMachineFunction(MachineFunction &MF) {
2741db9f3b2SDimitry Andric #ifndef NDEBUG
2751db9f3b2SDimitry Andric   // Make sure the tables are sorted.
2761db9f3b2SDimitry Andric   static std::atomic<bool> TableChecked(false);
2771db9f3b2SDimitry Andric   if (!TableChecked.load(std::memory_order_relaxed)) {
2781db9f3b2SDimitry Andric     assert(llvm::is_sorted(X86CompressEVEXTable) &&
2791db9f3b2SDimitry Andric            "X86CompressEVEXTable is not sorted!");
2801db9f3b2SDimitry Andric     TableChecked.store(true, std::memory_order_relaxed);
2811db9f3b2SDimitry Andric   }
2821db9f3b2SDimitry Andric #endif
2831db9f3b2SDimitry Andric   const X86Subtarget &ST = MF.getSubtarget<X86Subtarget>();
2841db9f3b2SDimitry Andric   if (!ST.hasAVX512() && !ST.hasEGPR() && !ST.hasNDD())
2851db9f3b2SDimitry Andric     return false;
2861db9f3b2SDimitry Andric 
2871db9f3b2SDimitry Andric   bool Changed = false;
2881db9f3b2SDimitry Andric 
2891db9f3b2SDimitry Andric   for (MachineBasicBlock &MBB : MF) {
2901db9f3b2SDimitry Andric     // Traverse the basic block.
2911db9f3b2SDimitry Andric     for (MachineInstr &MI : MBB)
2921db9f3b2SDimitry Andric       Changed |= CompressEVEXImpl(MI, ST);
2931db9f3b2SDimitry Andric   }
2941db9f3b2SDimitry Andric 
2951db9f3b2SDimitry Andric   return Changed;
2961db9f3b2SDimitry Andric }
2971db9f3b2SDimitry Andric 
2981db9f3b2SDimitry Andric INITIALIZE_PASS(CompressEVEXPass, COMP_EVEX_NAME, COMP_EVEX_DESC, false, false)
2991db9f3b2SDimitry Andric 
3001db9f3b2SDimitry Andric FunctionPass *llvm::createX86CompressEVEXPass() {
3011db9f3b2SDimitry Andric   return new CompressEVEXPass();
3021db9f3b2SDimitry Andric }
303