@@ -1954,21 +1954,17 @@ static const fltSemantics *getOpFltSemantics(uint8_t OperandType) {
1954
1954
// representation of the constant truncated to the 16 LSBs should be used.
1955
1955
case AMDGPU::OPERAND_REG_IMM_INT16:
1956
1956
case AMDGPU::OPERAND_REG_INLINE_C_INT16:
1957
- case AMDGPU::OPERAND_REG_INLINE_AC_INT16:
1958
1957
case AMDGPU::OPERAND_REG_IMM_INT32:
1959
1958
case AMDGPU::OPERAND_REG_IMM_FP32:
1960
1959
case AMDGPU::OPERAND_REG_IMM_FP32_DEFERRED:
1961
1960
case AMDGPU::OPERAND_REG_INLINE_C_INT32:
1962
1961
case AMDGPU::OPERAND_REG_INLINE_C_FP32:
1963
1962
case AMDGPU::OPERAND_REG_INLINE_AC_INT32:
1964
1963
case AMDGPU::OPERAND_REG_INLINE_AC_FP32:
1965
- case AMDGPU::OPERAND_REG_INLINE_C_V2FP32:
1966
1964
case AMDGPU::OPERAND_REG_IMM_V2FP32:
1967
- case AMDGPU::OPERAND_REG_INLINE_C_V2INT32:
1968
1965
case AMDGPU::OPERAND_REG_IMM_V2INT32:
1969
1966
case AMDGPU::OPERAND_REG_IMM_V2INT16:
1970
1967
case AMDGPU::OPERAND_REG_INLINE_C_V2INT16:
1971
- case AMDGPU::OPERAND_REG_INLINE_AC_V2INT16:
1972
1968
case AMDGPU::OPERAND_KIMM32:
1973
1969
case AMDGPU::OPERAND_INLINE_SPLIT_BARRIER_INT32:
1974
1970
return &APFloat::IEEEsingle ();
@@ -1982,17 +1978,13 @@ static const fltSemantics *getOpFltSemantics(uint8_t OperandType) {
1982
1978
case AMDGPU::OPERAND_REG_IMM_FP16_DEFERRED:
1983
1979
case AMDGPU::OPERAND_REG_INLINE_C_FP16:
1984
1980
case AMDGPU::OPERAND_REG_INLINE_C_V2FP16:
1985
- case AMDGPU::OPERAND_REG_INLINE_AC_FP16:
1986
- case AMDGPU::OPERAND_REG_INLINE_AC_V2FP16:
1987
1981
case AMDGPU::OPERAND_REG_IMM_V2FP16:
1988
1982
case AMDGPU::OPERAND_KIMM16:
1989
1983
return &APFloat::IEEEhalf ();
1990
1984
case AMDGPU::OPERAND_REG_IMM_BF16:
1991
1985
case AMDGPU::OPERAND_REG_IMM_BF16_DEFERRED:
1992
1986
case AMDGPU::OPERAND_REG_INLINE_C_BF16:
1993
1987
case AMDGPU::OPERAND_REG_INLINE_C_V2BF16:
1994
- case AMDGPU::OPERAND_REG_INLINE_AC_BF16:
1995
- case AMDGPU::OPERAND_REG_INLINE_AC_V2BF16:
1996
1988
case AMDGPU::OPERAND_REG_IMM_V2BF16:
1997
1989
return &APFloat::BFloat ();
1998
1990
default :
@@ -2315,8 +2307,6 @@ void AMDGPUOperand::addLiteralImmOperand(MCInst &Inst, int64_t Val, bool ApplyMo
2315
2307
case AMDGPU::OPERAND_REG_IMM_BF16_DEFERRED:
2316
2308
case AMDGPU::OPERAND_REG_INLINE_C_BF16:
2317
2309
case AMDGPU::OPERAND_REG_INLINE_C_V2BF16:
2318
- case AMDGPU::OPERAND_REG_INLINE_AC_BF16:
2319
- case AMDGPU::OPERAND_REG_INLINE_AC_V2BF16:
2320
2310
case AMDGPU::OPERAND_REG_IMM_V2BF16:
2321
2311
if (AsmParser->hasInv2PiInlineImm () && Literal == 0x3fc45f306725feed ) {
2322
2312
// This is the 1/(2*pi) which is going to be truncated to bf16 with the
@@ -2343,15 +2333,9 @@ void AMDGPUOperand::addLiteralImmOperand(MCInst &Inst, int64_t Val, bool ApplyMo
2343
2333
case AMDGPU::OPERAND_REG_INLINE_C_FP16:
2344
2334
case AMDGPU::OPERAND_REG_INLINE_C_V2INT16:
2345
2335
case AMDGPU::OPERAND_REG_INLINE_C_V2FP16:
2346
- case AMDGPU::OPERAND_REG_INLINE_AC_INT16:
2347
- case AMDGPU::OPERAND_REG_INLINE_AC_FP16:
2348
- case AMDGPU::OPERAND_REG_INLINE_AC_V2INT16:
2349
- case AMDGPU::OPERAND_REG_INLINE_AC_V2FP16:
2350
2336
case AMDGPU::OPERAND_REG_IMM_V2INT16:
2351
2337
case AMDGPU::OPERAND_REG_IMM_V2FP16:
2352
- case AMDGPU::OPERAND_REG_INLINE_C_V2FP32:
2353
2338
case AMDGPU::OPERAND_REG_IMM_V2FP32:
2354
- case AMDGPU::OPERAND_REG_INLINE_C_V2INT32:
2355
2339
case AMDGPU::OPERAND_REG_IMM_V2INT32:
2356
2340
case AMDGPU::OPERAND_KIMM32:
2357
2341
case AMDGPU::OPERAND_KIMM16:
@@ -2394,9 +2378,7 @@ void AMDGPUOperand::addLiteralImmOperand(MCInst &Inst, int64_t Val, bool ApplyMo
2394
2378
case AMDGPU::OPERAND_REG_IMM_V2BF16:
2395
2379
case AMDGPU::OPERAND_REG_IMM_V2FP16:
2396
2380
case AMDGPU::OPERAND_REG_IMM_V2FP32:
2397
- case AMDGPU::OPERAND_REG_INLINE_C_V2FP32:
2398
2381
case AMDGPU::OPERAND_REG_IMM_V2INT32:
2399
- case AMDGPU::OPERAND_REG_INLINE_C_V2INT32:
2400
2382
case AMDGPU::OPERAND_INLINE_SPLIT_BARRIER_INT32:
2401
2383
if (isSafeTruncation (Val, 32 ) &&
2402
2384
AMDGPU::isInlinableLiteral32 (static_cast <int32_t >(Val),
@@ -2430,7 +2412,6 @@ void AMDGPUOperand::addLiteralImmOperand(MCInst &Inst, int64_t Val, bool ApplyMo
2430
2412
2431
2413
case AMDGPU::OPERAND_REG_IMM_INT16:
2432
2414
case AMDGPU::OPERAND_REG_INLINE_C_INT16:
2433
- case AMDGPU::OPERAND_REG_INLINE_AC_INT16:
2434
2415
if (isSafeTruncation (Val, 16 ) &&
2435
2416
AMDGPU::isInlinableIntLiteral (static_cast <int16_t >(Val))) {
2436
2417
Inst.addOperand (MCOperand::createImm (Lo_32 (Val)));
@@ -2445,7 +2426,6 @@ void AMDGPUOperand::addLiteralImmOperand(MCInst &Inst, int64_t Val, bool ApplyMo
2445
2426
case AMDGPU::OPERAND_REG_INLINE_C_FP16:
2446
2427
case AMDGPU::OPERAND_REG_IMM_FP16:
2447
2428
case AMDGPU::OPERAND_REG_IMM_FP16_DEFERRED:
2448
- case AMDGPU::OPERAND_REG_INLINE_AC_FP16:
2449
2429
if (isSafeTruncation (Val, 16 ) &&
2450
2430
AMDGPU::isInlinableLiteralFP16 (static_cast <int16_t >(Val),
2451
2431
AsmParser->hasInv2PiInlineImm ())) {
@@ -2461,7 +2441,6 @@ void AMDGPUOperand::addLiteralImmOperand(MCInst &Inst, int64_t Val, bool ApplyMo
2461
2441
case AMDGPU::OPERAND_REG_IMM_BF16:
2462
2442
case AMDGPU::OPERAND_REG_IMM_BF16_DEFERRED:
2463
2443
case AMDGPU::OPERAND_REG_INLINE_C_BF16:
2464
- case AMDGPU::OPERAND_REG_INLINE_AC_BF16:
2465
2444
if (isSafeTruncation (Val, 16 ) &&
2466
2445
AMDGPU::isInlinableLiteralBF16 (static_cast <int16_t >(Val),
2467
2446
AsmParser->hasInv2PiInlineImm ())) {
@@ -2474,15 +2453,13 @@ void AMDGPUOperand::addLiteralImmOperand(MCInst &Inst, int64_t Val, bool ApplyMo
2474
2453
setImmKindLiteral ();
2475
2454
return ;
2476
2455
2477
- case AMDGPU::OPERAND_REG_INLINE_C_V2INT16:
2478
- case AMDGPU::OPERAND_REG_INLINE_AC_V2INT16: {
2456
+ case AMDGPU::OPERAND_REG_INLINE_C_V2INT16: {
2479
2457
assert (isSafeTruncation (Val, 16 ));
2480
2458
assert (AMDGPU::isInlinableIntLiteral (static_cast <int16_t >(Val)));
2481
2459
Inst.addOperand (MCOperand::createImm (Val));
2482
2460
return ;
2483
2461
}
2484
- case AMDGPU::OPERAND_REG_INLINE_C_V2FP16:
2485
- case AMDGPU::OPERAND_REG_INLINE_AC_V2FP16: {
2462
+ case AMDGPU::OPERAND_REG_INLINE_C_V2FP16: {
2486
2463
assert (isSafeTruncation (Val, 16 ));
2487
2464
assert (AMDGPU::isInlinableLiteralFP16 (static_cast <int16_t >(Val),
2488
2465
AsmParser->hasInv2PiInlineImm ()));
@@ -2491,8 +2468,7 @@ void AMDGPUOperand::addLiteralImmOperand(MCInst &Inst, int64_t Val, bool ApplyMo
2491
2468
return ;
2492
2469
}
2493
2470
2494
- case AMDGPU::OPERAND_REG_INLINE_C_V2BF16:
2495
- case AMDGPU::OPERAND_REG_INLINE_AC_V2BF16: {
2471
+ case AMDGPU::OPERAND_REG_INLINE_C_V2BF16: {
2496
2472
assert (isSafeTruncation (Val, 16 ));
2497
2473
assert (AMDGPU::isInlinableLiteralBF16 (static_cast <int16_t >(Val),
2498
2474
AsmParser->hasInv2PiInlineImm ()));
@@ -3623,34 +3599,28 @@ bool AMDGPUAsmParser::isInlineConstant(const MCInst &Inst,
3623
3599
case 2 : {
3624
3600
const unsigned OperandType = Desc.operands ()[OpIdx].OperandType ;
3625
3601
if (OperandType == AMDGPU::OPERAND_REG_IMM_INT16 ||
3626
- OperandType == AMDGPU::OPERAND_REG_INLINE_C_INT16 ||
3627
- OperandType == AMDGPU::OPERAND_REG_INLINE_AC_INT16)
3602
+ OperandType == AMDGPU::OPERAND_REG_INLINE_C_INT16)
3628
3603
return AMDGPU::isInlinableLiteralI16 (Val, hasInv2PiInlineImm ());
3629
3604
3630
3605
if (OperandType == AMDGPU::OPERAND_REG_INLINE_C_V2INT16 ||
3631
- OperandType == AMDGPU::OPERAND_REG_INLINE_AC_V2INT16 ||
3632
3606
OperandType == AMDGPU::OPERAND_REG_IMM_V2INT16)
3633
3607
return AMDGPU::isInlinableLiteralV2I16 (Val);
3634
3608
3635
3609
if (OperandType == AMDGPU::OPERAND_REG_INLINE_C_V2FP16 ||
3636
- OperandType == AMDGPU::OPERAND_REG_INLINE_AC_V2FP16 ||
3637
3610
OperandType == AMDGPU::OPERAND_REG_IMM_V2FP16)
3638
3611
return AMDGPU::isInlinableLiteralV2F16 (Val);
3639
3612
3640
3613
if (OperandType == AMDGPU::OPERAND_REG_INLINE_C_V2BF16 ||
3641
- OperandType == AMDGPU::OPERAND_REG_INLINE_AC_V2BF16 ||
3642
3614
OperandType == AMDGPU::OPERAND_REG_IMM_V2BF16)
3643
3615
return AMDGPU::isInlinableLiteralV2BF16 (Val);
3644
3616
3645
3617
if (OperandType == AMDGPU::OPERAND_REG_IMM_FP16 ||
3646
3618
OperandType == AMDGPU::OPERAND_REG_INLINE_C_FP16 ||
3647
- OperandType == AMDGPU::OPERAND_REG_INLINE_AC_FP16 ||
3648
3619
OperandType == AMDGPU::OPERAND_REG_IMM_FP16_DEFERRED)
3649
3620
return AMDGPU::isInlinableLiteralFP16 (Val, hasInv2PiInlineImm ());
3650
3621
3651
3622
if (OperandType == AMDGPU::OPERAND_REG_IMM_BF16 ||
3652
3623
OperandType == AMDGPU::OPERAND_REG_INLINE_C_BF16 ||
3653
- OperandType == AMDGPU::OPERAND_REG_INLINE_AC_BF16 ||
3654
3624
OperandType == AMDGPU::OPERAND_REG_IMM_BF16_DEFERRED)
3655
3625
return AMDGPU::isInlinableLiteralBF16 (Val, hasInv2PiInlineImm ());
3656
3626
0 commit comments