Commit | Line | Data |
---|---|---|
9bc89cd8 DW |
1 | /* |
2 | * Copyright © 2006, Intel Corporation. | |
3 | * | |
4 | * This program is free software; you can redistribute it and/or modify it | |
5 | * under the terms and conditions of the GNU General Public License, | |
6 | * version 2, as published by the Free Software Foundation. | |
7 | * | |
8 | * This program is distributed in the hope it will be useful, but WITHOUT | |
9 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
10 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for | |
11 | * more details. | |
12 | * | |
13 | * You should have received a copy of the GNU General Public License along with | |
14 | * this program; if not, write to the Free Software Foundation, Inc., | |
15 | * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA. | |
16 | * | |
17 | */ | |
18 | #ifndef _ASYNC_TX_H_ | |
19 | #define _ASYNC_TX_H_ | |
20 | #include <linux/dmaengine.h> | |
21 | #include <linux/spinlock.h> | |
22 | #include <linux/interrupt.h> | |
23 | ||
24 | /** | |
25 | * dma_chan_ref - object used to manage dma channels received from the | |
26 | * dmaengine core. | |
27 | * @chan - the channel being tracked | |
28 | * @node - node for the channel to be placed on async_tx_master_list | |
29 | * @rcu - for list_del_rcu | |
30 | * @count - number of times this channel is listed in the pool | |
31 | * (for channels with multiple capabiities) | |
32 | */ | |
33 | struct dma_chan_ref { | |
34 | struct dma_chan *chan; | |
35 | struct list_head node; | |
36 | struct rcu_head rcu; | |
37 | atomic_t count; | |
38 | }; | |
39 | ||
40 | /** | |
41 | * async_tx_flags - modifiers for the async_* calls | |
42 | * @ASYNC_TX_XOR_ZERO_DST: this flag must be used for xor operations where the | |
43 | * the destination address is not a source. The asynchronous case handles this | |
44 | * implicitly, the synchronous case needs to zero the destination block. | |
45 | * @ASYNC_TX_XOR_DROP_DST: this flag must be used if the destination address is | |
46 | * also one of the source addresses. In the synchronous case the destination | |
47 | * address is an implied source, whereas the asynchronous case it must be listed | |
48 | * as a source. The destination address must be the first address in the source | |
49 | * array. | |
50 | * @ASYNC_TX_ASSUME_COHERENT: skip cache maintenance operations | |
51 | * @ASYNC_TX_ACK: immediately ack the descriptor, precludes setting up a | |
52 | * dependency chain | |
53 | * @ASYNC_TX_DEP_ACK: ack the dependency descriptor. Useful for chaining. | |
54 | * @ASYNC_TX_KMAP_SRC: if the transaction is to be performed synchronously | |
55 | * take an atomic mapping (KM_USER0) on the source page(s) | |
56 | * @ASYNC_TX_KMAP_DST: if the transaction is to be performed synchronously | |
57 | * take an atomic mapping (KM_USER0) on the dest page(s) | |
58 | */ | |
59 | enum async_tx_flags { | |
60 | ASYNC_TX_XOR_ZERO_DST = (1 << 0), | |
61 | ASYNC_TX_XOR_DROP_DST = (1 << 1), | |
62 | ASYNC_TX_ASSUME_COHERENT = (1 << 2), | |
63 | ASYNC_TX_ACK = (1 << 3), | |
64 | ASYNC_TX_DEP_ACK = (1 << 4), | |
65 | ASYNC_TX_KMAP_SRC = (1 << 5), | |
66 | ASYNC_TX_KMAP_DST = (1 << 6), | |
67 | }; | |
68 | ||
69 | #ifdef CONFIG_DMA_ENGINE | |
70 | void async_tx_issue_pending_all(void); | |
71 | enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx); | |
72 | void async_tx_run_dependencies(struct dma_async_tx_descriptor *tx); | |
73 | struct dma_chan * | |
74 | async_tx_find_channel(struct dma_async_tx_descriptor *depend_tx, | |
75 | enum dma_transaction_type tx_type); | |
76 | #else | |
77 | static inline void async_tx_issue_pending_all(void) | |
78 | { | |
79 | do { } while (0); | |
80 | } | |
81 | ||
82 | static inline enum dma_status | |
83 | dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx) | |
84 | { | |
85 | return DMA_SUCCESS; | |
86 | } | |
87 | ||
88 | static inline void | |
89 | async_tx_run_dependencies(struct dma_async_tx_descriptor *tx, | |
90 | struct dma_chan *host_chan) | |
91 | { | |
92 | do { } while (0); | |
93 | } | |
94 | ||
95 | static inline struct dma_chan * | |
96 | async_tx_find_channel(struct dma_async_tx_descriptor *depend_tx, | |
97 | enum dma_transaction_type tx_type) | |
98 | { | |
99 | return NULL; | |
100 | } | |
101 | #endif | |
102 | ||
103 | /** | |
104 | * async_tx_sync_epilog - actions to take if an operation is run synchronously | |
105 | * @flags: async_tx flags | |
106 | * @depend_tx: transaction depends on depend_tx | |
107 | * @cb_fn: function to call when the transaction completes | |
108 | * @cb_fn_param: parameter to pass to the callback routine | |
109 | */ | |
110 | static inline void | |
111 | async_tx_sync_epilog(unsigned long flags, | |
112 | struct dma_async_tx_descriptor *depend_tx, | |
113 | dma_async_tx_callback cb_fn, void *cb_fn_param) | |
114 | { | |
115 | if (cb_fn) | |
116 | cb_fn(cb_fn_param); | |
117 | ||
118 | if (depend_tx && (flags & ASYNC_TX_DEP_ACK)) | |
119 | async_tx_ack(depend_tx); | |
120 | } | |
121 | ||
122 | void | |
123 | async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx, | |
124 | enum async_tx_flags flags, struct dma_async_tx_descriptor *depend_tx, | |
125 | dma_async_tx_callback cb_fn, void *cb_fn_param); | |
126 | ||
127 | struct dma_async_tx_descriptor * | |
128 | async_xor(struct page *dest, struct page **src_list, unsigned int offset, | |
129 | int src_cnt, size_t len, enum async_tx_flags flags, | |
130 | struct dma_async_tx_descriptor *depend_tx, | |
131 | dma_async_tx_callback cb_fn, void *cb_fn_param); | |
132 | ||
133 | struct dma_async_tx_descriptor * | |
134 | async_xor_zero_sum(struct page *dest, struct page **src_list, | |
135 | unsigned int offset, int src_cnt, size_t len, | |
136 | u32 *result, enum async_tx_flags flags, | |
137 | struct dma_async_tx_descriptor *depend_tx, | |
138 | dma_async_tx_callback cb_fn, void *cb_fn_param); | |
139 | ||
140 | struct dma_async_tx_descriptor * | |
141 | async_memcpy(struct page *dest, struct page *src, unsigned int dest_offset, | |
142 | unsigned int src_offset, size_t len, enum async_tx_flags flags, | |
143 | struct dma_async_tx_descriptor *depend_tx, | |
144 | dma_async_tx_callback cb_fn, void *cb_fn_param); | |
145 | ||
146 | struct dma_async_tx_descriptor * | |
147 | async_memset(struct page *dest, int val, unsigned int offset, | |
148 | size_t len, enum async_tx_flags flags, | |
149 | struct dma_async_tx_descriptor *depend_tx, | |
150 | dma_async_tx_callback cb_fn, void *cb_fn_param); | |
151 | ||
152 | struct dma_async_tx_descriptor * | |
153 | async_trigger_callback(enum async_tx_flags flags, | |
154 | struct dma_async_tx_descriptor *depend_tx, | |
155 | dma_async_tx_callback cb_fn, void *cb_fn_param); | |
156 | #endif /* _ASYNC_TX_H_ */ |