Commit | Line | Data |
---|---|---|
77c9a5da SR |
1 | #ifndef _FIREWIRE_CORE_H |
2 | #define _FIREWIRE_CORE_H | |
3 | ||
4 | #include <linux/dma-mapping.h> | |
5 | #include <linux/fs.h> | |
6 | #include <linux/list.h> | |
7 | #include <linux/idr.h> | |
8 | #include <linux/mm_types.h> | |
9 | #include <linux/rwsem.h> | |
10 | #include <linux/slab.h> | |
11 | #include <linux/types.h> | |
12 | ||
13 | #include <asm/atomic.h> | |
14 | ||
15 | struct device; | |
16 | struct fw_card; | |
17 | struct fw_device; | |
18 | struct fw_iso_buffer; | |
19 | struct fw_iso_context; | |
20 | struct fw_iso_packet; | |
21 | struct fw_node; | |
22 | struct fw_packet; | |
23 | ||
24 | ||
25 | /* -card */ | |
26 | ||
27 | /* bitfields within the PHY registers */ | |
28 | #define PHY_LINK_ACTIVE 0x80 | |
29 | #define PHY_CONTENDER 0x40 | |
30 | #define PHY_BUS_RESET 0x40 | |
31 | #define PHY_BUS_SHORT_RESET 0x40 | |
32 | ||
33 | #define BANDWIDTH_AVAILABLE_INITIAL 4915 | |
34 | #define BROADCAST_CHANNEL_INITIAL (1 << 31 | 31) | |
35 | #define BROADCAST_CHANNEL_VALID (1 << 30) | |
36 | ||
37 | struct fw_card_driver { | |
38 | /* | |
39 | * Enable the given card with the given initial config rom. | |
40 | * This function is expected to activate the card, and either | |
41 | * enable the PHY or set the link_on bit and initiate a bus | |
42 | * reset. | |
43 | */ | |
44 | int (*enable)(struct fw_card *card, u32 *config_rom, size_t length); | |
45 | ||
46 | int (*update_phy_reg)(struct fw_card *card, int address, | |
47 | int clear_bits, int set_bits); | |
48 | ||
49 | /* | |
50 | * Update the config rom for an enabled card. This function | |
51 | * should change the config rom that is presented on the bus | |
52 | * an initiate a bus reset. | |
53 | */ | |
54 | int (*set_config_rom)(struct fw_card *card, | |
55 | u32 *config_rom, size_t length); | |
56 | ||
57 | void (*send_request)(struct fw_card *card, struct fw_packet *packet); | |
58 | void (*send_response)(struct fw_card *card, struct fw_packet *packet); | |
59 | /* Calling cancel is valid once a packet has been submitted. */ | |
60 | int (*cancel_packet)(struct fw_card *card, struct fw_packet *packet); | |
61 | ||
62 | /* | |
63 | * Allow the specified node ID to do direct DMA out and in of | |
64 | * host memory. The card will disable this for all node when | |
65 | * a bus reset happens, so driver need to reenable this after | |
66 | * bus reset. Returns 0 on success, -ENODEV if the card | |
67 | * doesn't support this, -ESTALE if the generation doesn't | |
68 | * match. | |
69 | */ | |
70 | int (*enable_phys_dma)(struct fw_card *card, | |
71 | int node_id, int generation); | |
72 | ||
73 | u64 (*get_bus_time)(struct fw_card *card); | |
74 | ||
75 | struct fw_iso_context * | |
76 | (*allocate_iso_context)(struct fw_card *card, | |
77 | int type, int channel, size_t header_size); | |
78 | void (*free_iso_context)(struct fw_iso_context *ctx); | |
79 | ||
80 | int (*start_iso)(struct fw_iso_context *ctx, | |
81 | s32 cycle, u32 sync, u32 tags); | |
82 | ||
83 | int (*queue_iso)(struct fw_iso_context *ctx, | |
84 | struct fw_iso_packet *packet, | |
85 | struct fw_iso_buffer *buffer, | |
86 | unsigned long payload); | |
87 | ||
88 | int (*stop_iso)(struct fw_iso_context *ctx); | |
89 | }; | |
90 | ||
91 | void fw_card_initialize(struct fw_card *card, | |
92 | const struct fw_card_driver *driver, struct device *device); | |
93 | int fw_card_add(struct fw_card *card, | |
94 | u32 max_receive, u32 link_speed, u64 guid); | |
95 | void fw_core_remove_card(struct fw_card *card); | |
96 | int fw_core_initiate_bus_reset(struct fw_card *card, int short_reset); | |
97 | int fw_compute_block_crc(u32 *block); | |
98 | void fw_schedule_bm_work(struct fw_card *card, unsigned long delay); | |
99 | ||
100 | struct fw_descriptor { | |
101 | struct list_head link; | |
102 | size_t length; | |
103 | u32 immediate; | |
104 | u32 key; | |
105 | const u32 *data; | |
106 | }; | |
107 | ||
108 | int fw_core_add_descriptor(struct fw_descriptor *desc); | |
109 | void fw_core_remove_descriptor(struct fw_descriptor *desc); | |
110 | ||
111 | ||
112 | /* -cdev */ | |
113 | ||
114 | extern const struct file_operations fw_device_ops; | |
115 | ||
116 | void fw_device_cdev_update(struct fw_device *device); | |
117 | void fw_device_cdev_remove(struct fw_device *device); | |
118 | ||
119 | ||
120 | /* -device */ | |
121 | ||
122 | extern struct rw_semaphore fw_device_rwsem; | |
123 | extern struct idr fw_device_idr; | |
124 | extern int fw_cdev_major; | |
125 | ||
126 | struct fw_device *fw_device_get_by_devt(dev_t devt); | |
127 | void fw_device_set_broadcast_channel(struct fw_device *device, int generation); | |
128 | void fw_node_event(struct fw_card *card, struct fw_node *node, int event); | |
129 | ||
130 | ||
131 | /* -iso */ | |
132 | ||
133 | /* | |
134 | * The iso packet format allows for an immediate header/payload part | |
135 | * stored in 'header' immediately after the packet info plus an | |
136 | * indirect payload part that is pointer to by the 'payload' field. | |
137 | * Applications can use one or the other or both to implement simple | |
138 | * low-bandwidth streaming (e.g. audio) or more advanced | |
139 | * scatter-gather streaming (e.g. assembling video frame automatically). | |
140 | */ | |
141 | struct fw_iso_packet { | |
142 | u16 payload_length; /* Length of indirect payload. */ | |
143 | u32 interrupt:1; /* Generate interrupt on this packet */ | |
144 | u32 skip:1; /* Set to not send packet at all. */ | |
145 | u32 tag:2; | |
146 | u32 sy:4; | |
147 | u32 header_length:8; /* Length of immediate header. */ | |
148 | u32 header[0]; | |
149 | }; | |
150 | ||
151 | #define FW_ISO_CONTEXT_TRANSMIT 0 | |
152 | #define FW_ISO_CONTEXT_RECEIVE 1 | |
153 | ||
154 | #define FW_ISO_CONTEXT_MATCH_TAG0 1 | |
155 | #define FW_ISO_CONTEXT_MATCH_TAG1 2 | |
156 | #define FW_ISO_CONTEXT_MATCH_TAG2 4 | |
157 | #define FW_ISO_CONTEXT_MATCH_TAG3 8 | |
158 | #define FW_ISO_CONTEXT_MATCH_ALL_TAGS 15 | |
159 | ||
160 | /* | |
161 | * An iso buffer is just a set of pages mapped for DMA in the | |
162 | * specified direction. Since the pages are to be used for DMA, they | |
163 | * are not mapped into the kernel virtual address space. We store the | |
164 | * DMA address in the page private. The helper function | |
165 | * fw_iso_buffer_map() will map the pages into a given vma. | |
166 | */ | |
167 | struct fw_iso_buffer { | |
168 | enum dma_data_direction direction; | |
169 | struct page **pages; | |
170 | int page_count; | |
171 | }; | |
172 | ||
173 | typedef void (*fw_iso_callback_t)(struct fw_iso_context *context, | |
174 | u32 cycle, size_t header_length, | |
175 | void *header, void *data); | |
176 | ||
177 | struct fw_iso_context { | |
178 | struct fw_card *card; | |
179 | int type; | |
180 | int channel; | |
181 | int speed; | |
182 | size_t header_size; | |
183 | fw_iso_callback_t callback; | |
184 | void *callback_data; | |
185 | }; | |
186 | ||
187 | int fw_iso_buffer_init(struct fw_iso_buffer *buffer, struct fw_card *card, | |
188 | int page_count, enum dma_data_direction direction); | |
189 | int fw_iso_buffer_map(struct fw_iso_buffer *buffer, struct vm_area_struct *vma); | |
190 | void fw_iso_buffer_destroy(struct fw_iso_buffer *buffer, struct fw_card *card); | |
191 | ||
192 | struct fw_iso_context *fw_iso_context_create(struct fw_card *card, | |
193 | int type, int channel, int speed, size_t header_size, | |
194 | fw_iso_callback_t callback, void *callback_data); | |
195 | int fw_iso_context_queue(struct fw_iso_context *ctx, | |
196 | struct fw_iso_packet *packet, | |
197 | struct fw_iso_buffer *buffer, | |
198 | unsigned long payload); | |
199 | int fw_iso_context_start(struct fw_iso_context *ctx, | |
200 | int cycle, int sync, int tags); | |
201 | int fw_iso_context_stop(struct fw_iso_context *ctx); | |
202 | void fw_iso_context_destroy(struct fw_iso_context *ctx); | |
203 | ||
204 | void fw_iso_resource_manage(struct fw_card *card, int generation, | |
205 | u64 channels_mask, int *channel, int *bandwidth, bool allocate); | |
206 | ||
207 | ||
208 | /* -topology */ | |
209 | ||
210 | enum { | |
211 | FW_NODE_CREATED, | |
212 | FW_NODE_UPDATED, | |
213 | FW_NODE_DESTROYED, | |
214 | FW_NODE_LINK_ON, | |
215 | FW_NODE_LINK_OFF, | |
216 | FW_NODE_INITIATED_RESET, | |
217 | }; | |
218 | ||
219 | struct fw_node { | |
220 | u16 node_id; | |
221 | u8 color; | |
222 | u8 port_count; | |
223 | u8 link_on:1; | |
224 | u8 initiated_reset:1; | |
225 | u8 b_path:1; | |
226 | u8 phy_speed:2; /* As in the self ID packet. */ | |
227 | u8 max_speed:2; /* Minimum of all phy-speeds on the path from the | |
228 | * local node to this node. */ | |
229 | u8 max_depth:4; /* Maximum depth to any leaf node */ | |
230 | u8 max_hops:4; /* Max hops in this sub tree */ | |
231 | atomic_t ref_count; | |
232 | ||
233 | /* For serializing node topology into a list. */ | |
234 | struct list_head link; | |
235 | ||
236 | /* Upper layer specific data. */ | |
237 | void *data; | |
238 | ||
239 | struct fw_node *ports[0]; | |
240 | }; | |
241 | ||
242 | static inline struct fw_node *fw_node_get(struct fw_node *node) | |
243 | { | |
244 | atomic_inc(&node->ref_count); | |
245 | ||
246 | return node; | |
247 | } | |
248 | ||
249 | static inline void fw_node_put(struct fw_node *node) | |
250 | { | |
251 | if (atomic_dec_and_test(&node->ref_count)) | |
252 | kfree(node); | |
253 | } | |
254 | ||
255 | void fw_core_handle_bus_reset(struct fw_card *card, int node_id, | |
256 | int generation, int self_id_count, u32 *self_ids); | |
257 | void fw_destroy_nodes(struct fw_card *card); | |
258 | ||
259 | /* | |
260 | * Check whether new_generation is the immediate successor of old_generation. | |
261 | * Take counter roll-over at 255 (as per OHCI) into account. | |
262 | */ | |
263 | static inline bool is_next_generation(int new_generation, int old_generation) | |
264 | { | |
265 | return (new_generation & 0xff) == ((old_generation + 1) & 0xff); | |
266 | } | |
267 | ||
268 | ||
269 | /* -transaction */ | |
270 | ||
271 | #define TCODE_IS_READ_REQUEST(tcode) (((tcode) & ~1) == 4) | |
272 | #define TCODE_IS_BLOCK_PACKET(tcode) (((tcode) & 1) != 0) | |
273 | #define TCODE_IS_REQUEST(tcode) (((tcode) & 2) == 0) | |
274 | #define TCODE_IS_RESPONSE(tcode) (((tcode) & 2) != 0) | |
275 | #define TCODE_HAS_REQUEST_DATA(tcode) (((tcode) & 12) != 4) | |
276 | #define TCODE_HAS_RESPONSE_DATA(tcode) (((tcode) & 12) != 0) | |
277 | ||
278 | #define LOCAL_BUS 0xffc0 | |
279 | ||
280 | void fw_core_handle_request(struct fw_card *card, struct fw_packet *request); | |
281 | void fw_core_handle_response(struct fw_card *card, struct fw_packet *packet); | |
282 | void fw_fill_response(struct fw_packet *response, u32 *request_header, | |
283 | int rcode, void *payload, size_t length); | |
284 | void fw_flush_transactions(struct fw_card *card); | |
285 | void fw_send_phy_config(struct fw_card *card, | |
286 | int node_id, int generation, int gap_count); | |
287 | ||
288 | static inline int fw_stream_packet_destination_id(int tag, int channel, int sy) | |
289 | { | |
290 | return tag << 14 | channel << 8 | sy; | |
291 | } | |
292 | ||
293 | #endif /* _FIREWIRE_CORE_H */ |