@@ -284,48 +284,76 @@ define void @extract_v8i32_nxv16i32_8(<vscale x 16 x i32> %x, ptr %y) {
284
284
}
285
285
286
286
define void @extract_v8i1_v64i1_0 (ptr %x , ptr %y ) {
287
- ; CHECK-LABEL: extract_v8i1_v64i1_0:
288
- ; CHECK: # %bb.0:
289
- ; CHECK-NEXT: li a2, 64
290
- ; CHECK-NEXT: vsetvli zero, a2, e8, m4, ta, ma
291
- ; CHECK-NEXT: vlm.v v8, (a0)
292
- ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
293
- ; CHECK-NEXT: vsm.v v8, (a1)
294
- ; CHECK-NEXT: ret
287
+ ; CHECK-V-LABEL: extract_v8i1_v64i1_0:
288
+ ; CHECK-V: # %bb.0:
289
+ ; CHECK-V-NEXT: li a2, 64
290
+ ; CHECK-V-NEXT: vsetvli zero, a2, e8, m4, ta, ma
291
+ ; CHECK-V-NEXT: vlm.v v8, (a0)
292
+ ; CHECK-V-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
293
+ ; CHECK-V-NEXT: vsm.v v8, (a1)
294
+ ; CHECK-V-NEXT: ret
295
+ ;
296
+ ; CHECK-KNOWNVLEN128-LABEL: extract_v8i1_v64i1_0:
297
+ ; CHECK-KNOWNVLEN128: # %bb.0:
298
+ ; CHECK-KNOWNVLEN128-NEXT: vsetvli a2, zero, e8, m4, ta, ma
299
+ ; CHECK-KNOWNVLEN128-NEXT: vlm.v v8, (a0)
300
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
301
+ ; CHECK-KNOWNVLEN128-NEXT: vsm.v v8, (a1)
302
+ ; CHECK-KNOWNVLEN128-NEXT: ret
295
303
%a = load <64 x i1 >, ptr %x
296
304
%c = call <8 x i1 > @llvm.vector.extract.v8i1.v64i1 (<64 x i1 > %a , i64 0 )
297
305
store <8 x i1 > %c , ptr %y
298
306
ret void
299
307
}
300
308
301
309
define void @extract_v8i1_v64i1_8 (ptr %x , ptr %y ) {
302
- ; CHECK-LABEL: extract_v8i1_v64i1_8:
303
- ; CHECK: # %bb.0:
304
- ; CHECK-NEXT: li a2, 64
305
- ; CHECK-NEXT: vsetvli zero, a2, e8, m4, ta, ma
306
- ; CHECK-NEXT: vlm.v v8, (a0)
307
- ; CHECK-NEXT: vsetivli zero, 1, e8, mf2, ta, ma
308
- ; CHECK-NEXT: vslidedown.vi v8, v8, 1
309
- ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
310
- ; CHECK-NEXT: vsm.v v8, (a1)
311
- ; CHECK-NEXT: ret
310
+ ; CHECK-V-LABEL: extract_v8i1_v64i1_8:
311
+ ; CHECK-V: # %bb.0:
312
+ ; CHECK-V-NEXT: li a2, 64
313
+ ; CHECK-V-NEXT: vsetvli zero, a2, e8, m4, ta, ma
314
+ ; CHECK-V-NEXT: vlm.v v8, (a0)
315
+ ; CHECK-V-NEXT: vsetivli zero, 1, e8, mf2, ta, ma
316
+ ; CHECK-V-NEXT: vslidedown.vi v8, v8, 1
317
+ ; CHECK-V-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
318
+ ; CHECK-V-NEXT: vsm.v v8, (a1)
319
+ ; CHECK-V-NEXT: ret
320
+ ;
321
+ ; CHECK-KNOWNVLEN128-LABEL: extract_v8i1_v64i1_8:
322
+ ; CHECK-KNOWNVLEN128: # %bb.0:
323
+ ; CHECK-KNOWNVLEN128-NEXT: vsetvli a2, zero, e8, m4, ta, ma
324
+ ; CHECK-KNOWNVLEN128-NEXT: vlm.v v8, (a0)
325
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 1, e8, mf2, ta, ma
326
+ ; CHECK-KNOWNVLEN128-NEXT: vslidedown.vi v8, v8, 1
327
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
328
+ ; CHECK-KNOWNVLEN128-NEXT: vsm.v v8, (a1)
329
+ ; CHECK-KNOWNVLEN128-NEXT: ret
312
330
%a = load <64 x i1 >, ptr %x
313
331
%c = call <8 x i1 > @llvm.vector.extract.v8i1.v64i1 (<64 x i1 > %a , i64 8 )
314
332
store <8 x i1 > %c , ptr %y
315
333
ret void
316
334
}
317
335
318
336
define void @extract_v8i1_v64i1_48 (ptr %x , ptr %y ) {
319
- ; CHECK-LABEL: extract_v8i1_v64i1_48:
320
- ; CHECK: # %bb.0:
321
- ; CHECK-NEXT: li a2, 64
322
- ; CHECK-NEXT: vsetvli zero, a2, e8, m4, ta, ma
323
- ; CHECK-NEXT: vlm.v v8, (a0)
324
- ; CHECK-NEXT: vsetivli zero, 1, e8, mf2, ta, ma
325
- ; CHECK-NEXT: vslidedown.vi v8, v8, 6
326
- ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
327
- ; CHECK-NEXT: vsm.v v8, (a1)
328
- ; CHECK-NEXT: ret
337
+ ; CHECK-V-LABEL: extract_v8i1_v64i1_48:
338
+ ; CHECK-V: # %bb.0:
339
+ ; CHECK-V-NEXT: li a2, 64
340
+ ; CHECK-V-NEXT: vsetvli zero, a2, e8, m4, ta, ma
341
+ ; CHECK-V-NEXT: vlm.v v8, (a0)
342
+ ; CHECK-V-NEXT: vsetivli zero, 1, e8, mf2, ta, ma
343
+ ; CHECK-V-NEXT: vslidedown.vi v8, v8, 6
344
+ ; CHECK-V-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
345
+ ; CHECK-V-NEXT: vsm.v v8, (a1)
346
+ ; CHECK-V-NEXT: ret
347
+ ;
348
+ ; CHECK-KNOWNVLEN128-LABEL: extract_v8i1_v64i1_48:
349
+ ; CHECK-KNOWNVLEN128: # %bb.0:
350
+ ; CHECK-KNOWNVLEN128-NEXT: vsetvli a2, zero, e8, m4, ta, ma
351
+ ; CHECK-KNOWNVLEN128-NEXT: vlm.v v8, (a0)
352
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 1, e8, mf2, ta, ma
353
+ ; CHECK-KNOWNVLEN128-NEXT: vslidedown.vi v8, v8, 6
354
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
355
+ ; CHECK-KNOWNVLEN128-NEXT: vsm.v v8, (a1)
356
+ ; CHECK-KNOWNVLEN128-NEXT: ret
329
357
%a = load <64 x i1 >, ptr %x
330
358
%c = call <8 x i1 > @llvm.vector.extract.v8i1.v64i1 (<64 x i1 > %a , i64 48 )
331
359
store <8 x i1 > %c , ptr %y
@@ -407,79 +435,138 @@ define void @extract_v8i1_nxv64i1_192(<vscale x 64 x i1> %x, ptr %y) {
407
435
}
408
436
409
437
define void @extract_v2i1_v64i1_0 (ptr %x , ptr %y ) {
410
- ; CHECK-LABEL: extract_v2i1_v64i1_0:
411
- ; CHECK: # %bb.0:
412
- ; CHECK-NEXT: li a2, 64
413
- ; CHECK-NEXT: vsetvli zero, a2, e8, m4, ta, ma
414
- ; CHECK-NEXT: vlm.v v0, (a0)
415
- ; CHECK-NEXT: vsetivli zero, 2, e8, mf8, ta, ma
416
- ; CHECK-NEXT: vmv.v.i v8, 0
417
- ; CHECK-NEXT: vmerge.vim v8, v8, 1, v0
418
- ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
419
- ; CHECK-NEXT: vmv.v.i v9, 0
420
- ; CHECK-NEXT: vsetivli zero, 2, e8, mf2, tu, ma
421
- ; CHECK-NEXT: vmv.v.v v9, v8
422
- ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
423
- ; CHECK-NEXT: vmsne.vi v8, v9, 0
424
- ; CHECK-NEXT: vsm.v v8, (a1)
425
- ; CHECK-NEXT: ret
438
+ ; CHECK-V-LABEL: extract_v2i1_v64i1_0:
439
+ ; CHECK-V: # %bb.0:
440
+ ; CHECK-V-NEXT: li a2, 64
441
+ ; CHECK-V-NEXT: vsetvli zero, a2, e8, m4, ta, ma
442
+ ; CHECK-V-NEXT: vlm.v v0, (a0)
443
+ ; CHECK-V-NEXT: vsetivli zero, 2, e8, mf8, ta, ma
444
+ ; CHECK-V-NEXT: vmv.v.i v8, 0
445
+ ; CHECK-V-NEXT: vmerge.vim v8, v8, 1, v0
446
+ ; CHECK-V-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
447
+ ; CHECK-V-NEXT: vmv.v.i v9, 0
448
+ ; CHECK-V-NEXT: vsetivli zero, 2, e8, mf2, tu, ma
449
+ ; CHECK-V-NEXT: vmv.v.v v9, v8
450
+ ; CHECK-V-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
451
+ ; CHECK-V-NEXT: vmsne.vi v8, v9, 0
452
+ ; CHECK-V-NEXT: vsm.v v8, (a1)
453
+ ; CHECK-V-NEXT: ret
454
+ ;
455
+ ; CHECK-KNOWNVLEN128-LABEL: extract_v2i1_v64i1_0:
456
+ ; CHECK-KNOWNVLEN128: # %bb.0:
457
+ ; CHECK-KNOWNVLEN128-NEXT: vsetvli a2, zero, e8, m4, ta, ma
458
+ ; CHECK-KNOWNVLEN128-NEXT: vlm.v v0, (a0)
459
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 2, e8, mf8, ta, ma
460
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.i v8, 0
461
+ ; CHECK-KNOWNVLEN128-NEXT: vmerge.vim v8, v8, 1, v0
462
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
463
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.i v9, 0
464
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 2, e8, mf2, tu, ma
465
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.v v9, v8
466
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
467
+ ; CHECK-KNOWNVLEN128-NEXT: vmsne.vi v8, v9, 0
468
+ ; CHECK-KNOWNVLEN128-NEXT: vsm.v v8, (a1)
469
+ ; CHECK-KNOWNVLEN128-NEXT: ret
426
470
%a = load <64 x i1 >, ptr %x
427
471
%c = call <2 x i1 > @llvm.vector.extract.v2i1.v64i1 (<64 x i1 > %a , i64 0 )
428
472
store <2 x i1 > %c , ptr %y
429
473
ret void
430
474
}
431
475
432
476
define void @extract_v2i1_v64i1_2 (ptr %x , ptr %y ) {
433
- ; CHECK-LABEL: extract_v2i1_v64i1_2:
434
- ; CHECK: # %bb.0:
435
- ; CHECK-NEXT: li a2, 64
436
- ; CHECK-NEXT: vsetvli zero, a2, e8, m4, ta, ma
437
- ; CHECK-NEXT: vlm.v v0, (a0)
438
- ; CHECK-NEXT: vmv.v.i v8, 0
439
- ; CHECK-NEXT: vmerge.vim v8, v8, 1, v0
440
- ; CHECK-NEXT: vsetivli zero, 2, e8, m1, ta, ma
441
- ; CHECK-NEXT: vslidedown.vi v8, v8, 2
442
- ; CHECK-NEXT: vsetivli zero, 2, e8, mf8, ta, ma
443
- ; CHECK-NEXT: vmsne.vi v0, v8, 0
444
- ; CHECK-NEXT: vmv.v.i v8, 0
445
- ; CHECK-NEXT: vmerge.vim v8, v8, 1, v0
446
- ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
447
- ; CHECK-NEXT: vmv.v.i v9, 0
448
- ; CHECK-NEXT: vsetivli zero, 2, e8, mf2, tu, ma
449
- ; CHECK-NEXT: vmv.v.v v9, v8
450
- ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
451
- ; CHECK-NEXT: vmsne.vi v8, v9, 0
452
- ; CHECK-NEXT: vsm.v v8, (a1)
453
- ; CHECK-NEXT: ret
477
+ ; CHECK-V-LABEL: extract_v2i1_v64i1_2:
478
+ ; CHECK-V: # %bb.0:
479
+ ; CHECK-V-NEXT: li a2, 64
480
+ ; CHECK-V-NEXT: vsetvli zero, a2, e8, m4, ta, ma
481
+ ; CHECK-V-NEXT: vlm.v v0, (a0)
482
+ ; CHECK-V-NEXT: vmv.v.i v8, 0
483
+ ; CHECK-V-NEXT: vmerge.vim v8, v8, 1, v0
484
+ ; CHECK-V-NEXT: vsetivli zero, 2, e8, m1, ta, ma
485
+ ; CHECK-V-NEXT: vslidedown.vi v8, v8, 2
486
+ ; CHECK-V-NEXT: vsetivli zero, 2, e8, mf8, ta, ma
487
+ ; CHECK-V-NEXT: vmsne.vi v0, v8, 0
488
+ ; CHECK-V-NEXT: vmv.v.i v8, 0
489
+ ; CHECK-V-NEXT: vmerge.vim v8, v8, 1, v0
490
+ ; CHECK-V-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
491
+ ; CHECK-V-NEXT: vmv.v.i v9, 0
492
+ ; CHECK-V-NEXT: vsetivli zero, 2, e8, mf2, tu, ma
493
+ ; CHECK-V-NEXT: vmv.v.v v9, v8
494
+ ; CHECK-V-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
495
+ ; CHECK-V-NEXT: vmsne.vi v8, v9, 0
496
+ ; CHECK-V-NEXT: vsm.v v8, (a1)
497
+ ; CHECK-V-NEXT: ret
498
+ ;
499
+ ; CHECK-KNOWNVLEN128-LABEL: extract_v2i1_v64i1_2:
500
+ ; CHECK-KNOWNVLEN128: # %bb.0:
501
+ ; CHECK-KNOWNVLEN128-NEXT: vsetvli a2, zero, e8, m4, ta, ma
502
+ ; CHECK-KNOWNVLEN128-NEXT: vlm.v v0, (a0)
503
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.i v8, 0
504
+ ; CHECK-KNOWNVLEN128-NEXT: vmerge.vim v8, v8, 1, v0
505
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 2, e8, m1, ta, ma
506
+ ; CHECK-KNOWNVLEN128-NEXT: vslidedown.vi v8, v8, 2
507
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 2, e8, mf8, ta, ma
508
+ ; CHECK-KNOWNVLEN128-NEXT: vmsne.vi v0, v8, 0
509
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.i v8, 0
510
+ ; CHECK-KNOWNVLEN128-NEXT: vmerge.vim v8, v8, 1, v0
511
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
512
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.i v9, 0
513
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 2, e8, mf2, tu, ma
514
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.v v9, v8
515
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
516
+ ; CHECK-KNOWNVLEN128-NEXT: vmsne.vi v8, v9, 0
517
+ ; CHECK-KNOWNVLEN128-NEXT: vsm.v v8, (a1)
518
+ ; CHECK-KNOWNVLEN128-NEXT: ret
454
519
%a = load <64 x i1 >, ptr %x
455
520
%c = call <2 x i1 > @llvm.vector.extract.v2i1.v64i1 (<64 x i1 > %a , i64 2 )
456
521
store <2 x i1 > %c , ptr %y
457
522
ret void
458
523
}
459
524
460
525
define void @extract_v2i1_v64i1_42 (ptr %x , ptr %y ) {
461
- ; CHECK-LABEL: extract_v2i1_v64i1_42:
462
- ; CHECK: # %bb.0:
463
- ; CHECK-NEXT: li a2, 64
464
- ; CHECK-NEXT: vsetvli zero, a2, e8, m4, ta, ma
465
- ; CHECK-NEXT: vlm.v v0, (a0)
466
- ; CHECK-NEXT: vmv.v.i v8, 0
467
- ; CHECK-NEXT: vmerge.vim v8, v8, 1, v0
468
- ; CHECK-NEXT: li a0, 42
469
- ; CHECK-NEXT: vsetivli zero, 2, e8, m4, ta, ma
470
- ; CHECK-NEXT: vslidedown.vx v8, v8, a0
471
- ; CHECK-NEXT: vsetivli zero, 2, e8, mf8, ta, ma
472
- ; CHECK-NEXT: vmsne.vi v0, v8, 0
473
- ; CHECK-NEXT: vmv.v.i v8, 0
474
- ; CHECK-NEXT: vmerge.vim v8, v8, 1, v0
475
- ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
476
- ; CHECK-NEXT: vmv.v.i v9, 0
477
- ; CHECK-NEXT: vsetivli zero, 2, e8, mf2, tu, ma
478
- ; CHECK-NEXT: vmv.v.v v9, v8
479
- ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
480
- ; CHECK-NEXT: vmsne.vi v8, v9, 0
481
- ; CHECK-NEXT: vsm.v v8, (a1)
482
- ; CHECK-NEXT: ret
526
+ ; CHECK-V-LABEL: extract_v2i1_v64i1_42:
527
+ ; CHECK-V: # %bb.0:
528
+ ; CHECK-V-NEXT: li a2, 64
529
+ ; CHECK-V-NEXT: vsetvli zero, a2, e8, m4, ta, ma
530
+ ; CHECK-V-NEXT: vlm.v v0, (a0)
531
+ ; CHECK-V-NEXT: vmv.v.i v8, 0
532
+ ; CHECK-V-NEXT: vmerge.vim v8, v8, 1, v0
533
+ ; CHECK-V-NEXT: li a0, 42
534
+ ; CHECK-V-NEXT: vsetivli zero, 2, e8, m4, ta, ma
535
+ ; CHECK-V-NEXT: vslidedown.vx v8, v8, a0
536
+ ; CHECK-V-NEXT: vsetivli zero, 2, e8, mf8, ta, ma
537
+ ; CHECK-V-NEXT: vmsne.vi v0, v8, 0
538
+ ; CHECK-V-NEXT: vmv.v.i v8, 0
539
+ ; CHECK-V-NEXT: vmerge.vim v8, v8, 1, v0
540
+ ; CHECK-V-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
541
+ ; CHECK-V-NEXT: vmv.v.i v9, 0
542
+ ; CHECK-V-NEXT: vsetivli zero, 2, e8, mf2, tu, ma
543
+ ; CHECK-V-NEXT: vmv.v.v v9, v8
544
+ ; CHECK-V-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
545
+ ; CHECK-V-NEXT: vmsne.vi v8, v9, 0
546
+ ; CHECK-V-NEXT: vsm.v v8, (a1)
547
+ ; CHECK-V-NEXT: ret
548
+ ;
549
+ ; CHECK-KNOWNVLEN128-LABEL: extract_v2i1_v64i1_42:
550
+ ; CHECK-KNOWNVLEN128: # %bb.0:
551
+ ; CHECK-KNOWNVLEN128-NEXT: vsetvli a2, zero, e8, m4, ta, ma
552
+ ; CHECK-KNOWNVLEN128-NEXT: vlm.v v0, (a0)
553
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.i v8, 0
554
+ ; CHECK-KNOWNVLEN128-NEXT: vmerge.vim v8, v8, 1, v0
555
+ ; CHECK-KNOWNVLEN128-NEXT: li a0, 42
556
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 2, e8, m4, ta, ma
557
+ ; CHECK-KNOWNVLEN128-NEXT: vslidedown.vx v8, v8, a0
558
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 2, e8, mf8, ta, ma
559
+ ; CHECK-KNOWNVLEN128-NEXT: vmsne.vi v0, v8, 0
560
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.i v8, 0
561
+ ; CHECK-KNOWNVLEN128-NEXT: vmerge.vim v8, v8, 1, v0
562
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
563
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.i v9, 0
564
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 2, e8, mf2, tu, ma
565
+ ; CHECK-KNOWNVLEN128-NEXT: vmv.v.v v9, v8
566
+ ; CHECK-KNOWNVLEN128-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
567
+ ; CHECK-KNOWNVLEN128-NEXT: vmsne.vi v8, v9, 0
568
+ ; CHECK-KNOWNVLEN128-NEXT: vsm.v v8, (a1)
569
+ ; CHECK-KNOWNVLEN128-NEXT: ret
483
570
%a = load <64 x i1 >, ptr %x
484
571
%c = call <2 x i1 > @llvm.vector.extract.v2i1.v64i1 (<64 x i1 > %a , i64 42 )
485
572
store <2 x i1 > %c , ptr %y
@@ -660,6 +747,3 @@ declare <2 x i8> @llvm.vector.extract.v2i8.nxv2i8(<vscale x 2 x i8> %vec, i64 %i
660
747
661
748
declare <2 x i32 > @llvm.vector.extract.v2i32.nxv16i32 (<vscale x 16 x i32 > %vec , i64 %idx )
662
749
declare <8 x i32 > @llvm.vector.extract.v8i32.nxv16i32 (<vscale x 16 x i32 > %vec , i64 %idx )
663
- ;; NOTE: These prefixes are unused and the list is autogenerated. Do not add tests below this line:
664
- ; CHECK-KNOWNVLEN128: {{.*}}
665
- ; CHECK-V: {{.*}}
0 commit comments