swiotlb: remove various exports
All these symbols are only used by arch dma_ops implementations or xen-swiotlb. None of which can be modular. Signed-off-by: Christoph Hellwig <hch@lst.de> Acked-by: Christian König <christian.koenig@amd.com>
This commit is contained in:
parent
0176adb004
commit
4bd89ed39b
1 changed files with 0 additions and 13 deletions
|
@ -605,7 +605,6 @@ phys_addr_t swiotlb_tbl_map_single(struct device *hwdev,
|
||||||
|
|
||||||
return tlb_addr;
|
return tlb_addr;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(swiotlb_tbl_map_single);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Allocates bounce buffer and returns its kernel virtual address.
|
* Allocates bounce buffer and returns its kernel virtual address.
|
||||||
|
@ -675,7 +674,6 @@ void swiotlb_tbl_unmap_single(struct device *hwdev, phys_addr_t tlb_addr,
|
||||||
}
|
}
|
||||||
spin_unlock_irqrestore(&io_tlb_lock, flags);
|
spin_unlock_irqrestore(&io_tlb_lock, flags);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(swiotlb_tbl_unmap_single);
|
|
||||||
|
|
||||||
void swiotlb_tbl_sync_single(struct device *hwdev, phys_addr_t tlb_addr,
|
void swiotlb_tbl_sync_single(struct device *hwdev, phys_addr_t tlb_addr,
|
||||||
size_t size, enum dma_data_direction dir,
|
size_t size, enum dma_data_direction dir,
|
||||||
|
@ -707,7 +705,6 @@ void swiotlb_tbl_sync_single(struct device *hwdev, phys_addr_t tlb_addr,
|
||||||
BUG();
|
BUG();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(swiotlb_tbl_sync_single);
|
|
||||||
|
|
||||||
static inline bool dma_coherent_ok(struct device *dev, dma_addr_t addr,
|
static inline bool dma_coherent_ok(struct device *dev, dma_addr_t addr,
|
||||||
size_t size)
|
size_t size)
|
||||||
|
@ -884,7 +881,6 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct page *page,
|
||||||
|
|
||||||
return swiotlb_phys_to_dma(dev, io_tlb_overflow_buffer);
|
return swiotlb_phys_to_dma(dev, io_tlb_overflow_buffer);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(swiotlb_map_page);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Unmap a single streaming mode DMA translation. The dma_addr and size must
|
* Unmap a single streaming mode DMA translation. The dma_addr and size must
|
||||||
|
@ -925,7 +921,6 @@ void swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr,
|
||||||
{
|
{
|
||||||
unmap_single(hwdev, dev_addr, size, dir, attrs);
|
unmap_single(hwdev, dev_addr, size, dir, attrs);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(swiotlb_unmap_page);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Make physical memory consistent for a single streaming mode DMA translation
|
* Make physical memory consistent for a single streaming mode DMA translation
|
||||||
|
@ -963,7 +958,6 @@ swiotlb_sync_single_for_cpu(struct device *hwdev, dma_addr_t dev_addr,
|
||||||
{
|
{
|
||||||
swiotlb_sync_single(hwdev, dev_addr, size, dir, SYNC_FOR_CPU);
|
swiotlb_sync_single(hwdev, dev_addr, size, dir, SYNC_FOR_CPU);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(swiotlb_sync_single_for_cpu);
|
|
||||||
|
|
||||||
void
|
void
|
||||||
swiotlb_sync_single_for_device(struct device *hwdev, dma_addr_t dev_addr,
|
swiotlb_sync_single_for_device(struct device *hwdev, dma_addr_t dev_addr,
|
||||||
|
@ -971,7 +965,6 @@ swiotlb_sync_single_for_device(struct device *hwdev, dma_addr_t dev_addr,
|
||||||
{
|
{
|
||||||
swiotlb_sync_single(hwdev, dev_addr, size, dir, SYNC_FOR_DEVICE);
|
swiotlb_sync_single(hwdev, dev_addr, size, dir, SYNC_FOR_DEVICE);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(swiotlb_sync_single_for_device);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Map a set of buffers described by scatterlist in streaming mode for DMA.
|
* Map a set of buffers described by scatterlist in streaming mode for DMA.
|
||||||
|
@ -1023,7 +1016,6 @@ swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, int nelems,
|
||||||
}
|
}
|
||||||
return nelems;
|
return nelems;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(swiotlb_map_sg_attrs);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Unmap a set of streaming mode DMA translations. Again, cpu read rules
|
* Unmap a set of streaming mode DMA translations. Again, cpu read rules
|
||||||
|
@ -1043,7 +1035,6 @@ swiotlb_unmap_sg_attrs(struct device *hwdev, struct scatterlist *sgl,
|
||||||
unmap_single(hwdev, sg->dma_address, sg_dma_len(sg), dir,
|
unmap_single(hwdev, sg->dma_address, sg_dma_len(sg), dir,
|
||||||
attrs);
|
attrs);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(swiotlb_unmap_sg_attrs);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Make physical memory consistent for a set of streaming mode DMA translations
|
* Make physical memory consistent for a set of streaming mode DMA translations
|
||||||
|
@ -1071,7 +1062,6 @@ swiotlb_sync_sg_for_cpu(struct device *hwdev, struct scatterlist *sg,
|
||||||
{
|
{
|
||||||
swiotlb_sync_sg(hwdev, sg, nelems, dir, SYNC_FOR_CPU);
|
swiotlb_sync_sg(hwdev, sg, nelems, dir, SYNC_FOR_CPU);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(swiotlb_sync_sg_for_cpu);
|
|
||||||
|
|
||||||
void
|
void
|
||||||
swiotlb_sync_sg_for_device(struct device *hwdev, struct scatterlist *sg,
|
swiotlb_sync_sg_for_device(struct device *hwdev, struct scatterlist *sg,
|
||||||
|
@ -1079,14 +1069,12 @@ swiotlb_sync_sg_for_device(struct device *hwdev, struct scatterlist *sg,
|
||||||
{
|
{
|
||||||
swiotlb_sync_sg(hwdev, sg, nelems, dir, SYNC_FOR_DEVICE);
|
swiotlb_sync_sg(hwdev, sg, nelems, dir, SYNC_FOR_DEVICE);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(swiotlb_sync_sg_for_device);
|
|
||||||
|
|
||||||
int
|
int
|
||||||
swiotlb_dma_mapping_error(struct device *hwdev, dma_addr_t dma_addr)
|
swiotlb_dma_mapping_error(struct device *hwdev, dma_addr_t dma_addr)
|
||||||
{
|
{
|
||||||
return (dma_addr == swiotlb_phys_to_dma(hwdev, io_tlb_overflow_buffer));
|
return (dma_addr == swiotlb_phys_to_dma(hwdev, io_tlb_overflow_buffer));
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(swiotlb_dma_mapping_error);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Return whether the given device DMA address mask can be supported
|
* Return whether the given device DMA address mask can be supported
|
||||||
|
@ -1099,7 +1087,6 @@ swiotlb_dma_supported(struct device *hwdev, u64 mask)
|
||||||
{
|
{
|
||||||
return swiotlb_phys_to_dma(hwdev, io_tlb_end - 1) <= mask;
|
return swiotlb_phys_to_dma(hwdev, io_tlb_end - 1) <= mask;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(swiotlb_dma_supported);
|
|
||||||
|
|
||||||
#ifdef CONFIG_DMA_DIRECT_OPS
|
#ifdef CONFIG_DMA_DIRECT_OPS
|
||||||
void *swiotlb_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle,
|
void *swiotlb_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle,
|
||||||
|
|
Loading…
Reference in a new issue