1
2
3
4
5
6
7
8
9
10
11
12
13
14
15#ifndef _ASM_TILE_DMA_MAPPING_H
16#define _ASM_TILE_DMA_MAPPING_H
17
18#include <linux/mm.h>
19#include <linux/scatterlist.h>
20#include <linux/cache.h>
21#include <linux/io.h>
22
23#ifdef __tilegx__
24#define ARCH_HAS_DMA_GET_REQUIRED_MASK
25#endif
26
27extern const struct dma_map_ops *tile_dma_map_ops;
28extern const struct dma_map_ops *gx_pci_dma_map_ops;
29extern const struct dma_map_ops *gx_legacy_pci_dma_map_ops;
30extern const struct dma_map_ops *gx_hybrid_pci_dma_map_ops;
31
32static inline const struct dma_map_ops *get_arch_dma_ops(struct bus_type *bus)
33{
34 return tile_dma_map_ops;
35}
36
37static inline dma_addr_t get_dma_offset(struct device *dev)
38{
39 return dev->archdata.dma_offset;
40}
41
42static inline void set_dma_offset(struct device *dev, dma_addr_t off)
43{
44 dev->archdata.dma_offset = off;
45}
46
47static inline dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
48{
49 return paddr;
50}
51
52static inline phys_addr_t dma_to_phys(struct device *dev, dma_addr_t daddr)
53{
54 return daddr;
55}
56
57static inline void dma_mark_clean(void *addr, size_t size) {}
58
59static inline bool dma_capable(struct device *dev, dma_addr_t addr, size_t size)
60{
61 if (!dev->dma_mask)
62 return 0;
63
64 return addr + size - 1 <= *dev->dma_mask;
65}
66
67#define HAVE_ARCH_DMA_SET_MASK 1
68int dma_set_mask(struct device *dev, u64 mask);
69
70
71
72
73
74static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size,
75 enum dma_data_direction direction)
76{
77}
78
79#endif
80