@@ -605,7 +605,6 @@ phys_addr_t swiotlb_tbl_map_single(struct device *hwdev,
605
605
606
606
return tlb_addr ;
607
607
}
608
- EXPORT_SYMBOL_GPL (swiotlb_tbl_map_single );
609
608
610
609
/*
611
610
* Allocates bounce buffer and returns its kernel virtual address.
@@ -675,7 +674,6 @@ void swiotlb_tbl_unmap_single(struct device *hwdev, phys_addr_t tlb_addr,
675
674
}
676
675
spin_unlock_irqrestore (& io_tlb_lock , flags );
677
676
}
678
- EXPORT_SYMBOL_GPL (swiotlb_tbl_unmap_single );
679
677
680
678
void swiotlb_tbl_sync_single (struct device * hwdev , phys_addr_t tlb_addr ,
681
679
size_t size , enum dma_data_direction dir ,
@@ -707,7 +705,6 @@ void swiotlb_tbl_sync_single(struct device *hwdev, phys_addr_t tlb_addr,
707
705
BUG ();
708
706
}
709
707
}
710
- EXPORT_SYMBOL_GPL (swiotlb_tbl_sync_single );
711
708
712
709
static inline bool dma_coherent_ok (struct device * dev , dma_addr_t addr ,
713
710
size_t size )
@@ -884,7 +881,6 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct page *page,
884
881
885
882
return swiotlb_phys_to_dma (dev , io_tlb_overflow_buffer );
886
883
}
887
- EXPORT_SYMBOL_GPL (swiotlb_map_page );
888
884
889
885
/*
890
886
* Unmap a single streaming mode DMA translation. The dma_addr and size must
@@ -925,7 +921,6 @@ void swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr,
925
921
{
926
922
unmap_single (hwdev , dev_addr , size , dir , attrs );
927
923
}
928
- EXPORT_SYMBOL_GPL (swiotlb_unmap_page );
929
924
930
925
/*
931
926
* Make physical memory consistent for a single streaming mode DMA translation
@@ -963,15 +958,13 @@ swiotlb_sync_single_for_cpu(struct device *hwdev, dma_addr_t dev_addr,
963
958
{
964
959
swiotlb_sync_single (hwdev , dev_addr , size , dir , SYNC_FOR_CPU );
965
960
}
966
- EXPORT_SYMBOL (swiotlb_sync_single_for_cpu );
967
961
968
962
void
969
963
swiotlb_sync_single_for_device (struct device * hwdev , dma_addr_t dev_addr ,
970
964
size_t size , enum dma_data_direction dir )
971
965
{
972
966
swiotlb_sync_single (hwdev , dev_addr , size , dir , SYNC_FOR_DEVICE );
973
967
}
974
- EXPORT_SYMBOL (swiotlb_sync_single_for_device );
975
968
976
969
/*
977
970
* Map a set of buffers described by scatterlist in streaming mode for DMA.
@@ -1023,7 +1016,6 @@ swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, int nelems,
1023
1016
}
1024
1017
return nelems ;
1025
1018
}
1026
- EXPORT_SYMBOL (swiotlb_map_sg_attrs );
1027
1019
1028
1020
/*
1029
1021
* Unmap a set of streaming mode DMA translations. Again, cpu read rules
@@ -1043,7 +1035,6 @@ swiotlb_unmap_sg_attrs(struct device *hwdev, struct scatterlist *sgl,
1043
1035
unmap_single (hwdev , sg -> dma_address , sg_dma_len (sg ), dir ,
1044
1036
attrs );
1045
1037
}
1046
- EXPORT_SYMBOL (swiotlb_unmap_sg_attrs );
1047
1038
1048
1039
/*
1049
1040
* Make physical memory consistent for a set of streaming mode DMA translations
@@ -1071,22 +1062,19 @@ swiotlb_sync_sg_for_cpu(struct device *hwdev, struct scatterlist *sg,
1071
1062
{
1072
1063
swiotlb_sync_sg (hwdev , sg , nelems , dir , SYNC_FOR_CPU );
1073
1064
}
1074
- EXPORT_SYMBOL (swiotlb_sync_sg_for_cpu );
1075
1065
1076
1066
void
1077
1067
swiotlb_sync_sg_for_device (struct device * hwdev , struct scatterlist * sg ,
1078
1068
int nelems , enum dma_data_direction dir )
1079
1069
{
1080
1070
swiotlb_sync_sg (hwdev , sg , nelems , dir , SYNC_FOR_DEVICE );
1081
1071
}
1082
- EXPORT_SYMBOL (swiotlb_sync_sg_for_device );
1083
1072
1084
1073
int
1085
1074
swiotlb_dma_mapping_error (struct device * hwdev , dma_addr_t dma_addr )
1086
1075
{
1087
1076
return (dma_addr == swiotlb_phys_to_dma (hwdev , io_tlb_overflow_buffer ));
1088
1077
}
1089
- EXPORT_SYMBOL (swiotlb_dma_mapping_error );
1090
1078
1091
1079
/*
1092
1080
* Return whether the given device DMA address mask can be supported
@@ -1099,7 +1087,6 @@ swiotlb_dma_supported(struct device *hwdev, u64 mask)
1099
1087
{
1100
1088
return swiotlb_phys_to_dma (hwdev , io_tlb_end - 1 ) <= mask ;
1101
1089
}
1102
- EXPORT_SYMBOL (swiotlb_dma_supported );
1103
1090
1104
1091
#ifdef CONFIG_DMA_DIRECT_OPS
1105
1092
void * swiotlb_alloc (struct device * dev , size_t size , dma_addr_t * dma_handle ,
0 commit comments