@@ -300,6 +300,7 @@ pci_map_single(struct pci_dev *pdev, void *cpu_addr, size_t size, int dir)
300
300
dac_allowed = pdev ? pci_dac_dma_supported (pdev , pdev -> dma_mask ) : 0 ;
301
301
return pci_map_single_1 (pdev , cpu_addr , size , dac_allowed );
302
302
}
303
+ EXPORT_SYMBOL (pci_map_single );
303
304
304
305
dma_addr_t
305
306
pci_map_page (struct pci_dev * pdev , struct page * page , unsigned long offset ,
@@ -314,6 +315,7 @@ pci_map_page(struct pci_dev *pdev, struct page *page, unsigned long offset,
314
315
return pci_map_single_1 (pdev , (char * )page_address (page ) + offset ,
315
316
size , dac_allowed );
316
317
}
318
+ EXPORT_SYMBOL (pci_map_page );
317
319
318
320
/* Unmap a single streaming mode DMA translation. The DMA_ADDR and
319
321
SIZE must match what was provided for in a previous pci_map_single
@@ -379,13 +381,15 @@ pci_unmap_single(struct pci_dev *pdev, dma_addr_t dma_addr, size_t size,
379
381
DBGA2 ("pci_unmap_single: sg [%lx,%lx] np %ld from %p\n" ,
380
382
dma_addr , size , npages , __builtin_return_address (0 ));
381
383
}
384
+ EXPORT_SYMBOL (pci_unmap_single );
382
385
383
386
void
384
387
pci_unmap_page (struct pci_dev * pdev , dma_addr_t dma_addr ,
385
388
size_t size , int direction )
386
389
{
387
390
pci_unmap_single (pdev , dma_addr , size , direction );
388
391
}
392
+ EXPORT_SYMBOL (pci_unmap_page );
389
393
390
394
/* Allocate and map kernel buffer using consistent mode DMA for PCI
391
395
device. Returns non-NULL cpu-view pointer to the buffer if
@@ -427,6 +431,7 @@ pci_alloc_consistent(struct pci_dev *pdev, size_t size, dma_addr_t *dma_addrp)
427
431
428
432
return cpu_addr ;
429
433
}
434
+ EXPORT_SYMBOL (pci_alloc_consistent );
430
435
431
436
/* Free and unmap a consistent DMA buffer. CPU_ADDR and DMA_ADDR must
432
437
be values that were returned from pci_alloc_consistent. SIZE must
@@ -444,7 +449,7 @@ pci_free_consistent(struct pci_dev *pdev, size_t size, void *cpu_addr,
444
449
DBGA2 ("pci_free_consistent: [%x,%lx] from %p\n" ,
445
450
dma_addr , size , __builtin_return_address (0 ));
446
451
}
447
-
452
+ EXPORT_SYMBOL ( pci_free_consistent );
448
453
449
454
/* Classify the elements of the scatterlist. Write dma_address
450
455
of each element with:
@@ -672,6 +677,7 @@ pci_map_sg(struct pci_dev *pdev, struct scatterlist *sg, int nents,
672
677
pci_unmap_sg (pdev , start , out - start , direction );
673
678
return 0 ;
674
679
}
680
+ EXPORT_SYMBOL (pci_map_sg );
675
681
676
682
/* Unmap a set of streaming mode DMA translations. Again, cpu read
677
683
rules concerning calls here are the same as for pci_unmap_single()
@@ -752,6 +758,7 @@ pci_unmap_sg(struct pci_dev *pdev, struct scatterlist *sg, int nents,
752
758
753
759
DBGA ("pci_unmap_sg: %ld entries\n" , nents - (end - sg ));
754
760
}
761
+ EXPORT_SYMBOL (pci_unmap_sg );
755
762
756
763
757
764
/* Return whether the given PCI device DMA address mask can be
@@ -786,6 +793,7 @@ pci_dma_supported(struct pci_dev *pdev, u64 mask)
786
793
787
794
return 0 ;
788
795
}
796
+ EXPORT_SYMBOL (pci_dma_supported );
789
797
790
798
791
799
/*
@@ -908,6 +916,7 @@ pci_dac_dma_supported(struct pci_dev *dev, u64 mask)
908
916
909
917
return ok ;
910
918
}
919
+ EXPORT_SYMBOL (pci_dac_dma_supported );
911
920
912
921
dma64_addr_t
913
922
pci_dac_page_to_dma (struct pci_dev * pdev , struct page * page ,
@@ -917,20 +926,22 @@ pci_dac_page_to_dma(struct pci_dev *pdev, struct page *page,
917
926
+ __pa (page_address (page ))
918
927
+ (dma64_addr_t ) offset );
919
928
}
929
+ EXPORT_SYMBOL (pci_dac_page_to_dma );
920
930
921
931
struct page *
922
932
pci_dac_dma_to_page (struct pci_dev * pdev , dma64_addr_t dma_addr )
923
933
{
924
934
unsigned long paddr = (dma_addr & PAGE_MASK ) - alpha_mv .pci_dac_offset ;
925
935
return virt_to_page (__va (paddr ));
926
936
}
937
+ EXPORT_SYMBOL (pci_dac_dma_to_page );
927
938
928
939
unsigned long
929
940
pci_dac_dma_to_offset (struct pci_dev * pdev , dma64_addr_t dma_addr )
930
941
{
931
942
return (dma_addr & ~PAGE_MASK );
932
943
}
933
-
944
+ EXPORT_SYMBOL ( pci_dac_dma_to_offset );
934
945
935
946
/* Helper for generic DMA-mapping functions. */
936
947
@@ -957,6 +968,7 @@ alpha_gendev_to_pci(struct device *dev)
957
968
/* This assumes ISA bus master with dma_mask 0xffffff. */
958
969
return NULL ;
959
970
}
971
+ EXPORT_SYMBOL (alpha_gendev_to_pci );
960
972
961
973
int
962
974
dma_set_mask (struct device * dev , u64 mask )
@@ -969,3 +981,4 @@ dma_set_mask(struct device *dev, u64 mask)
969
981
970
982
return 0 ;
971
983
}
984
+ EXPORT_SYMBOL (dma_set_mask );
0 commit comments