@@ -237,7 +237,7 @@ void upscale_f32(const float *x, float *dst, const int nb00, const int nb01,
237
237
int i02 = i12 / sf2;
238
238
int i03 = i13 / sf3;
239
239
240
- dst[index ] = *(float *)((char *)x + i03 * nb03 + i02 * nb02 + i01 * nb01 + i00 * nb00);
240
+ dst[index ] = *(const float *)((const char *)x + i03 * nb03 + i02 * nb02 + i01 * nb01 + i00 * nb00);
241
241
}
242
242
243
243
void pad_f32 (const float *x, float *dst, const int ne0, const int ne00, const int ne01, const int ne02,
@@ -251,8 +251,7 @@ void pad_f32(const float *x, float *dst, const int ne0, const int ne00, const i
251
251
// operation
252
252
int offset_dst = nidx + item_ct1.get_group (1 ) * ne0 +
253
253
item_ct1.get_group (0 ) * ne0 * item_ct1.get_group_range (1 );
254
- if (nidx < ne00 && item_ct1.get_group (1 ) < ne01 &&
255
- item_ct1.get_group (0 ) < ne02) {
254
+ if (nidx < ne00 && item_ct1.get_group (1 ) < (size_t ) ne01 && item_ct1.get_group (0 ) < (size_t ) ne02) {
256
255
int offset_src = nidx + item_ct1.get_group (1 ) * ne00 +
257
256
item_ct1.get_group (0 ) * ne00 * ne01;
258
257
dst[offset_dst] = x[offset_src];
@@ -520,9 +519,10 @@ inline void ggml_sycl_op_silu(ggml_backend_sycl_context & ctx, const ggml_tensor
520
519
521
520
silu_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
522
521
523
- (void ) src1;
524
- (void ) dst;
525
- (void ) src1_dd;
522
+ GGML_UNUSED (src1);
523
+ GGML_UNUSED (dst);
524
+ GGML_UNUSED (src1_dd);
525
+ GGML_UNUSED (ctx);
526
526
}
527
527
528
528
inline void ggml_sycl_op_gelu (ggml_backend_sycl_context & ctx, const ggml_tensor *src0, const ggml_tensor *src1,
@@ -535,9 +535,10 @@ inline void ggml_sycl_op_gelu(ggml_backend_sycl_context & ctx, const ggml_tensor
535
535
536
536
gelu_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
537
537
538
- (void ) src1;
539
- (void ) dst;
540
- (void ) src1_dd;
538
+ GGML_UNUSED (src1);
539
+ GGML_UNUSED (dst);
540
+ GGML_UNUSED (src1_dd);
541
+ GGML_UNUSED (ctx);
541
542
}
542
543
inline void ggml_sycl_op_gelu_quick (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
543
544
const ggml_tensor *src1, ggml_tensor *dst,
@@ -550,9 +551,10 @@ inline void ggml_sycl_op_gelu_quick(ggml_backend_sycl_context & ctx, const ggml_
550
551
551
552
gelu_quick_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
552
553
553
- (void ) src1;
554
- (void ) dst;
555
- (void ) src1_dd;
554
+ GGML_UNUSED (src1);
555
+ GGML_UNUSED (dst);
556
+ GGML_UNUSED (src1_dd);
557
+ GGML_UNUSED (ctx);
556
558
}
557
559
558
560
inline void ggml_sycl_op_tanh (ggml_backend_sycl_context & ctx, const ggml_tensor *src0, const ggml_tensor *src1,
@@ -564,9 +566,10 @@ inline void ggml_sycl_op_tanh(ggml_backend_sycl_context & ctx, const ggml_tensor
564
566
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
565
567
tanh_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
566
568
567
- (void ) src1;
568
- (void ) dst;
569
- (void ) src1_dd;
569
+ GGML_UNUSED (src1);
570
+ GGML_UNUSED (dst);
571
+ GGML_UNUSED (src1_dd);
572
+ GGML_UNUSED (ctx);
570
573
}
571
574
572
575
inline void ggml_sycl_op_relu (ggml_backend_sycl_context & ctx, const ggml_tensor *src0, const ggml_tensor *src1,
@@ -579,9 +582,10 @@ inline void ggml_sycl_op_relu(ggml_backend_sycl_context & ctx, const ggml_tensor
579
582
580
583
relu_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
581
584
582
- (void ) src1;
583
- (void ) dst;
584
- (void ) src1_dd;
585
+ GGML_UNUSED (src1);
586
+ GGML_UNUSED (dst);
587
+ GGML_UNUSED (src1_dd);
588
+ GGML_UNUSED (ctx);
585
589
}
586
590
587
591
inline void ggml_sycl_op_hardsigmoid (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -595,9 +599,10 @@ inline void ggml_sycl_op_hardsigmoid(ggml_backend_sycl_context & ctx, const ggml
595
599
596
600
hardsigmoid_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
597
601
598
- (void ) src1;
599
- (void ) dst;
600
- (void ) src1_dd;
602
+ GGML_UNUSED (src1);
603
+ GGML_UNUSED (dst);
604
+ GGML_UNUSED (src1_dd);
605
+ GGML_UNUSED (ctx);
601
606
}
602
607
603
608
inline void ggml_sycl_op_hardswish (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -610,9 +615,10 @@ inline void ggml_sycl_op_hardswish(ggml_backend_sycl_context & ctx, const ggml_t
610
615
611
616
hardswish_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
612
617
613
- (void ) src1;
614
- (void ) dst;
615
- (void ) src1_dd;
618
+ GGML_UNUSED (src1);
619
+ GGML_UNUSED (dst);
620
+ GGML_UNUSED (src1_dd);
621
+ GGML_UNUSED (ctx);
616
622
}
617
623
618
624
inline void ggml_sycl_op_exp (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -625,9 +631,10 @@ inline void ggml_sycl_op_exp(ggml_backend_sycl_context & ctx, const ggml_tensor
625
631
626
632
exp_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
627
633
628
- (void ) src1;
629
- (void ) dst;
630
- (void ) src1_dd;
634
+ GGML_UNUSED (src1);
635
+ GGML_UNUSED (dst);
636
+ GGML_UNUSED (src1_dd);
637
+ GGML_UNUSED (ctx);
631
638
}
632
639
633
640
inline void ggml_sycl_op_log (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -640,9 +647,10 @@ inline void ggml_sycl_op_log(ggml_backend_sycl_context & ctx, const ggml_tensor
640
647
641
648
log_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
642
649
643
- (void ) src1;
644
- (void ) dst;
645
- (void ) src1_dd;
650
+ GGML_UNUSED (src1);
651
+ GGML_UNUSED (dst);
652
+ GGML_UNUSED (src1_dd);
653
+ GGML_UNUSED (ctx);
646
654
}
647
655
648
656
inline void ggml_sycl_op_sigmoid (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -655,9 +663,10 @@ inline void ggml_sycl_op_sigmoid(ggml_backend_sycl_context & ctx, const ggml_ten
655
663
656
664
sigmoid_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
657
665
658
- (void ) src1;
659
- (void ) dst;
660
- (void ) src1_dd;
666
+ GGML_UNUSED (src1);
667
+ GGML_UNUSED (dst);
668
+ GGML_UNUSED (src1_dd);
669
+ GGML_UNUSED (ctx);
661
670
}
662
671
663
672
inline void ggml_sycl_op_sqrt (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -670,9 +679,10 @@ inline void ggml_sycl_op_sqrt(ggml_backend_sycl_context & ctx, const ggml_tensor
670
679
671
680
sqrt_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
672
681
673
- (void ) src1;
674
- (void ) dst;
675
- (void ) src1_dd;
682
+ GGML_UNUSED (src1);
683
+ GGML_UNUSED (dst);
684
+ GGML_UNUSED (src1_dd);
685
+ GGML_UNUSED (ctx);
676
686
}
677
687
678
688
inline void ggml_sycl_op_sin (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -685,9 +695,10 @@ inline void ggml_sycl_op_sin(ggml_backend_sycl_context & ctx, const ggml_tensor
685
695
686
696
sin_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
687
697
688
- (void ) src1;
689
- (void ) dst;
690
- (void ) src1_dd;
698
+ GGML_UNUSED (src1);
699
+ GGML_UNUSED (dst);
700
+ GGML_UNUSED (src1_dd);
701
+ GGML_UNUSED (ctx);
691
702
}
692
703
693
704
inline void ggml_sycl_op_cos (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -700,9 +711,10 @@ inline void ggml_sycl_op_cos(ggml_backend_sycl_context & ctx, const ggml_tensor
700
711
701
712
cos_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
702
713
703
- (void ) src1;
704
- (void ) dst;
705
- (void ) src1_dd;
714
+ GGML_UNUSED (src1);
715
+ GGML_UNUSED (dst);
716
+ GGML_UNUSED (src1_dd);
717
+ GGML_UNUSED (ctx);
706
718
}
707
719
708
720
inline void ggml_sycl_op_step (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -715,9 +727,10 @@ inline void ggml_sycl_op_step(ggml_backend_sycl_context & ctx, const ggml_tensor
715
727
716
728
step_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
717
729
718
- (void ) src1;
719
- (void ) dst;
720
- (void ) src1_dd;
730
+ GGML_UNUSED (src1);
731
+ GGML_UNUSED (dst);
732
+ GGML_UNUSED (src1_dd);
733
+ GGML_UNUSED (ctx);
721
734
}
722
735
723
736
inline void ggml_sycl_op_neg (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -730,9 +743,10 @@ inline void ggml_sycl_op_neg(ggml_backend_sycl_context & ctx, const ggml_tensor
730
743
731
744
neg_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
732
745
733
- (void ) src1;
734
- (void ) dst;
735
- (void ) src1_dd;
746
+ GGML_UNUSED (src1);
747
+ GGML_UNUSED (dst);
748
+ GGML_UNUSED (src1_dd);
749
+ GGML_UNUSED (ctx);
736
750
}
737
751
738
752
inline void ggml_sycl_op_leaky_relu (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -749,9 +763,10 @@ inline void ggml_sycl_op_leaky_relu(ggml_backend_sycl_context & ctx, const ggml_
749
763
750
764
leaky_relu_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), negative_slope, main_stream);
751
765
752
- (void ) src1;
753
- (void ) dst;
754
- (void ) src1_dd;
766
+ GGML_UNUSED (src1);
767
+ GGML_UNUSED (dst);
768
+ GGML_UNUSED (src1_dd);
769
+ GGML_UNUSED (ctx);
755
770
}
756
771
757
772
inline void ggml_sycl_op_sqr (ggml_backend_sycl_context & ctx, const ggml_tensor *src0, const ggml_tensor *src1,
@@ -764,9 +779,10 @@ inline void ggml_sycl_op_sqr(ggml_backend_sycl_context & ctx, const ggml_tensor
764
779
765
780
sqr_f32_sycl (src0_dd, dst_dd, ggml_nelements (src0), main_stream);
766
781
767
- (void ) src1;
768
- (void ) dst;
769
- (void ) src1_dd;
782
+ GGML_UNUSED (src1);
783
+ GGML_UNUSED (dst);
784
+ GGML_UNUSED (src1_dd);
785
+ GGML_UNUSED (ctx);
770
786
}
771
787
772
788
inline void ggml_sycl_op_upscale (ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
@@ -787,9 +803,10 @@ inline void ggml_sycl_op_upscale(ggml_backend_sycl_context & ctx, const ggml_ten
787
803
dst->ne [0 ], dst->ne [1 ], dst->ne [2 ], dst->ne [3 ], sf0, sf1, sf2, sf3,
788
804
main_stream);
789
805
790
- (void ) src1;
791
- (void ) dst;
792
- (void ) src1_dd;
806
+ GGML_UNUSED (src1);
807
+ GGML_UNUSED (dst);
808
+ GGML_UNUSED (src1_dd);
809
+ GGML_UNUSED (ctx);
793
810
}
794
811
795
812
inline void ggml_sycl_op_pad (ggml_backend_sycl_context & ctx, const ggml_tensor *src0, const ggml_tensor *src1,
@@ -805,9 +822,10 @@ inline void ggml_sycl_op_pad(ggml_backend_sycl_context & ctx, const ggml_tensor
805
822
src0->ne [0 ], src0->ne [1 ], src0->ne [2 ],
806
823
dst->ne [0 ], dst->ne [1 ], dst->ne [2 ], main_stream);
807
824
808
- (void ) src1;
809
- (void ) dst;
810
- (void ) src1_dd;
825
+ GGML_UNUSED (src1);
826
+ GGML_UNUSED (dst);
827
+ GGML_UNUSED (src1_dd);
828
+ GGML_UNUSED (ctx);
811
829
}
812
830
813
831
inline void ggml_sycl_op_acc (ggml_backend_sycl_context & ctx, const ggml_tensor *src0, const ggml_tensor *src1,
@@ -827,7 +845,8 @@ inline void ggml_sycl_op_acc(ggml_backend_sycl_context & ctx, const ggml_tensor
827
845
828
846
acc_f32_sycl (src0_dd, src1_dd, dst_dd, ggml_nelements (dst), src1->ne [0 ], src1->ne [1 ], src1->ne [2 ], nb1, nb2, offset, main_stream);
829
847
830
- (void ) dst;
848
+ GGML_UNUSED (dst);
849
+ GGML_UNUSED (ctx);
831
850
}
832
851
833
852
inline void ggml_sycl_op_add (ggml_backend_sycl_context & ctx, const ggml_tensor *src0, const ggml_tensor *src1,
0 commit comments