@@ -103,8 +103,7 @@ AArch64LegalizerInfo::AArch64LegalizerInfo(const AArch64Subtarget &ST)
103
103
getActionDefinitionsBuilder (G_BSWAP)
104
104
.legalFor ({s32, s64, v4s32, v2s32, v2s64})
105
105
.clampScalar (0 , s32, s64)
106
- .widenScalarToNextPow2 (0 )
107
- .customIf (typeIs (0 , v2s16)); // custom lower as G_REV32 + G_LSHR
106
+ .widenScalarToNextPow2 (0 );
108
107
109
108
getActionDefinitionsBuilder ({G_ADD, G_SUB, G_MUL, G_AND, G_OR, G_XOR})
110
109
.legalFor ({s32, s64, v2s32, v4s32, v4s16, v8s16, v16s8, v8s8})
@@ -799,8 +798,6 @@ bool AArch64LegalizerInfo::legalizeCustom(LegalizerHelper &Helper,
799
798
case TargetOpcode::G_LOAD:
800
799
case TargetOpcode::G_STORE:
801
800
return legalizeLoadStore (MI, MRI, MIRBuilder, Observer);
802
- case TargetOpcode::G_BSWAP:
803
- return legalizeBSwap (MI, MRI, MIRBuilder);
804
801
case TargetOpcode::G_SHL:
805
802
case TargetOpcode::G_ASHR:
806
803
case TargetOpcode::G_LSHR:
@@ -1015,46 +1012,6 @@ bool AArch64LegalizerInfo::legalizeLoadStore(
1015
1012
return true ;
1016
1013
}
1017
1014
1018
- bool AArch64LegalizerInfo::legalizeBSwap (MachineInstr &MI,
1019
- MachineRegisterInfo &MRI,
1020
- MachineIRBuilder &MIRBuilder) const {
1021
- assert (MI.getOpcode () == TargetOpcode::G_BSWAP);
1022
-
1023
- // The <2 x half> case needs special lowering because there isn't an
1024
- // instruction that does that directly. Instead, we widen to <8 x i8>
1025
- // and emit a G_REV32 followed by a G_LSHR knowing that instruction selection
1026
- // will later match them as:
1027
- //
1028
- // rev32.8b v0, v0
1029
- // ushr.2s v0, v0, #16
1030
- //
1031
- // We could emit those here directly, but it seems better to keep things as
1032
- // generic as possible through legalization, and avoid committing layering
1033
- // violations by legalizing & selecting here at the same time.
1034
-
1035
- Register ValReg = MI.getOperand (1 ).getReg ();
1036
- assert (LLT::fixed_vector (2 , 16 ) == MRI.getType (ValReg));
1037
- const LLT v2s32 = LLT::fixed_vector (2 , 32 );
1038
- const LLT v8s8 = LLT::fixed_vector (8 , 8 );
1039
- const LLT s32 = LLT::scalar (32 );
1040
-
1041
- auto Undef = MIRBuilder.buildUndef (v8s8);
1042
- auto Insert =
1043
- MIRBuilder
1044
- .buildInstr (TargetOpcode::INSERT_SUBREG, {v8s8}, {Undef, ValReg})
1045
- .addImm (AArch64::ssub);
1046
- auto Rev32 = MIRBuilder.buildInstr (AArch64::G_REV32, {v8s8}, {Insert});
1047
- auto Bitcast = MIRBuilder.buildBitcast (v2s32, Rev32);
1048
- auto Amt = MIRBuilder.buildConstant (v2s32, 16 );
1049
- auto UShr =
1050
- MIRBuilder.buildInstr (TargetOpcode::G_LSHR, {v2s32}, {Bitcast, Amt});
1051
- auto Zero = MIRBuilder.buildConstant (s32, 0 );
1052
- auto Extract = MIRBuilder.buildExtractVectorElement (s32, UShr, Zero);
1053
- MIRBuilder.buildBitcast ({MI.getOperand (0 ).getReg ()}, Extract);
1054
- MI.eraseFromParent ();
1055
- return true ;
1056
- }
1057
-
1058
1015
bool AArch64LegalizerInfo::legalizeVaArg (MachineInstr &MI,
1059
1016
MachineRegisterInfo &MRI,
1060
1017
MachineIRBuilder &MIRBuilder) const {
0 commit comments