@@ -564,21 +564,21 @@ static void ggml_gemv_q4_0_4x4_q8_0(int n, float * GGML_RESTRICT s, size_t bs, c
564
564
565
565
#if ! ((defined(_MSC_VER)) && ! defined(__clang__)) && defined(__aarch64__) && defined(__ARM_NEON) && defined(__ARM_FEATURE_DOTPROD)
566
566
if (ggml_cpu_has_neon () && ggml_cpu_has_dotprod ()) {
567
- const block_q4_0x4 * b_ptr = (const block_q4_0x4 *)vx;
567
+ const block_q4_0x4 * b_ptr = (const block_q4_0x4 *) vx;
568
568
569
569
for (int c = 0 ; c < nc; c += ncols_interleaved) {
570
- const block_q8_0 * a_ptr = (const block_q8_0 *)vy;
570
+ const block_q8_0 * a_ptr = (const block_q8_0 *) vy;
571
571
float32x4_t acc = vdupq_n_f32 (0 );
572
572
for (int b = 0 ; b < nb; b++) {
573
- int8x16_t b0 = vld1q_s8 ((const int8_t *)b_ptr->qs );
574
- int8x16_t b1 = vld1q_s8 ((const int8_t *)b_ptr->qs + 16 );
575
- int8x16_t b2 = vld1q_s8 ((const int8_t *)b_ptr->qs + 32 );
576
- int8x16_t b3 = vld1q_s8 ((const int8_t *)b_ptr->qs + 48 );
577
- float16x4_t bd = vld1_f16 ((const __fp16 *)b_ptr->d );
573
+ int8x16_t b0 = vld1q_s8 ((const int8_t *) b_ptr->qs );
574
+ int8x16_t b1 = vld1q_s8 ((const int8_t *) b_ptr->qs + 16 );
575
+ int8x16_t b2 = vld1q_s8 ((const int8_t *) b_ptr->qs + 32 );
576
+ int8x16_t b3 = vld1q_s8 ((const int8_t *) b_ptr->qs + 48 );
577
+ float16x4_t bd = vld1_f16 ((const __fp16 *) b_ptr->d );
578
578
579
579
int8x16_t a0 = vld1q_s8 (a_ptr->qs );
580
580
int8x16_t a1 = vld1q_s8 (a_ptr->qs + qk/2 );
581
- float16x4_t ad = vld1_dup_f16 ((const __fp16 *)&a_ptr->d );
581
+ float16x4_t ad = vld1_dup_f16 ((const __fp16 *) &a_ptr->d );
582
582
583
583
int32x4_t ret = vdupq_n_s32 (0 );
584
584
@@ -649,23 +649,23 @@ static void ggml_gemv_q4_0_4x8_q8_0(int n, float * GGML_RESTRICT s, size_t bs, c
649
649
650
650
#if ! ((defined(_MSC_VER)) && ! defined(__clang__)) && defined(__aarch64__) && defined(__ARM_NEON) && defined(__ARM_FEATURE_DOTPROD)
651
651
if (ggml_cpu_has_neon () && ggml_cpu_has_dotprod ()) {
652
- const block_q4_0x4 * b_ptr = (const block_q4_0x4 *)vx;
652
+ const block_q4_0x4 * b_ptr = (const block_q4_0x4 *) vx;
653
653
654
654
for (int c = 0 ; c < nc; c += ncols_interleaved) {
655
- const block_q8_0 * a_ptr = (const block_q8_0 *)vy;
655
+ const block_q8_0 * a_ptr = (const block_q8_0 *) vy;
656
656
float32x4_t acc = vdupq_n_f32 (0 );
657
657
for (int b = 0 ; b < nb; b++) {
658
- int8x16_t b0 = vld1q_s8 ((const int8_t *)b_ptr->qs );
659
- int8x16_t b1 = vld1q_s8 ((const int8_t *)b_ptr->qs + 16 );
660
- int8x16_t b2 = vld1q_s8 ((const int8_t *)b_ptr->qs + 32 );
661
- int8x16_t b3 = vld1q_s8 ((const int8_t *)b_ptr->qs + 48 );
662
- float16x4_t bd = vld1_f16 ((const __fp16 *)b_ptr->d );
663
-
664
- int8x16_t a0 = (int8x16_t )vld1q_dup_s64 ((const int64_t *)a_ptr->qs );
665
- int8x16_t a1 = (int8x16_t )vld1q_dup_s64 ((const int64_t *)a_ptr->qs + 1 );
666
- int8x16_t a2 = (int8x16_t )vld1q_dup_s64 ((const int64_t *)a_ptr->qs + 2 );
667
- int8x16_t a3 = (int8x16_t )vld1q_dup_s64 ((const int64_t *)a_ptr->qs + 3 );
668
- float16x4_t ad = vld1_dup_f16 ((const __fp16 *)&a_ptr->d );
658
+ int8x16_t b0 = vld1q_s8 ((const int8_t *) b_ptr->qs );
659
+ int8x16_t b1 = vld1q_s8 ((const int8_t *) b_ptr->qs + 16 );
660
+ int8x16_t b2 = vld1q_s8 ((const int8_t *) b_ptr->qs + 32 );
661
+ int8x16_t b3 = vld1q_s8 ((const int8_t *) b_ptr->qs + 48 );
662
+ float16x4_t bd = vld1_f16 ((const __fp16 *) b_ptr->d );
663
+
664
+ int8x16_t a0 = (int8x16_t ) vld1q_dup_s64 ((const int64_t *) a_ptr->qs );
665
+ int8x16_t a1 = (int8x16_t ) vld1q_dup_s64 ((const int64_t *) a_ptr->qs + 1 );
666
+ int8x16_t a2 = (int8x16_t ) vld1q_dup_s64 ((const int64_t *) a_ptr->qs + 2 );
667
+ int8x16_t a3 = (int8x16_t ) vld1q_dup_s64 ((const int64_t *) a_ptr->qs + 3 );
668
+ float16x4_t ad = vld1_dup_f16 ((const __fp16 *) &a_ptr->d );
669
669
670
670
int32x4_t ret0 = vdupq_n_s32 (0 );
671
671
int32x4_t ret1 = vdupq_n_s32 (0 );
0 commit comments