Commit | Line | Data |
---|---|---|
cee9fbd8 | 1 | /* QLogic qedr NIC Driver |
e8f1cb50 | 2 | * Copyright (c) 2015-2017 QLogic Corporation |
cee9fbd8 RA |
3 | * |
4 | * This software is available to you under a choice of one of two | |
5 | * licenses. You may choose to be licensed under the terms of the GNU | |
6 | * General Public License (GPL) Version 2, available from the file | |
7 | * COPYING in the main directory of this source tree, or the | |
8 | * OpenIB.org BSD license below: | |
9 | * | |
10 | * Redistribution and use in source and binary forms, with or | |
11 | * without modification, are permitted provided that the following | |
12 | * conditions are met: | |
13 | * | |
14 | * - Redistributions of source code must retain the above | |
15 | * copyright notice, this list of conditions and the following | |
16 | * disclaimer. | |
17 | * | |
18 | * - Redistributions in binary form must reproduce the above | |
19 | * copyright notice, this list of conditions and the following | |
20 | * disclaimer in the documentation and /or other materials | |
21 | * provided with the distribution. | |
22 | * | |
23 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | |
24 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | |
25 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND | |
26 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS | |
27 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN | |
28 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN | |
29 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | |
30 | * SOFTWARE. | |
31 | */ | |
32 | #include <linux/pci.h> | |
33 | #include <linux/netdevice.h> | |
34 | #include <linux/list.h> | |
35 | #include <linux/mutex.h> | |
b262a06e | 36 | #include <linux/qed/qede_rdma.h> |
cee9fbd8 RA |
37 | #include "qede.h" |
38 | ||
39 | static struct qedr_driver *qedr_drv; | |
40 | static LIST_HEAD(qedr_dev_list); | |
41 | static DEFINE_MUTEX(qedr_dev_list_lock); | |
42 | ||
bbfcd1e8 | 43 | bool qede_rdma_supported(struct qede_dev *dev) |
cee9fbd8 RA |
44 | { |
45 | return dev->dev_info.common.rdma_supported; | |
46 | } | |
47 | ||
bbfcd1e8 | 48 | static void _qede_rdma_dev_add(struct qede_dev *edev) |
cee9fbd8 RA |
49 | { |
50 | if (!qedr_drv) | |
51 | return; | |
52 | ||
ccc67ef5 TT |
53 | /* Leftovers from previous error recovery */ |
54 | edev->rdma_info.exp_recovery = false; | |
cee9fbd8 RA |
55 | edev->rdma_info.qedr_dev = qedr_drv->add(edev->cdev, edev->pdev, |
56 | edev->ndev); | |
57 | } | |
58 | ||
bbfcd1e8 | 59 | static int qede_rdma_create_wq(struct qede_dev *edev) |
cee9fbd8 | 60 | { |
bbfcd1e8 | 61 | INIT_LIST_HEAD(&edev->rdma_info.rdma_event_list); |
af6565ad MK |
62 | kref_init(&edev->rdma_info.refcnt); |
63 | init_completion(&edev->rdma_info.event_comp); | |
64 | ||
bbfcd1e8 MK |
65 | edev->rdma_info.rdma_wq = create_singlethread_workqueue("rdma_wq"); |
66 | if (!edev->rdma_info.rdma_wq) { | |
cee9fbd8 RA |
67 | DP_NOTICE(edev, "qedr: Could not create workqueue\n"); |
68 | return -ENOMEM; | |
69 | } | |
70 | ||
71 | return 0; | |
72 | } | |
73 | ||
bbfcd1e8 | 74 | static void qede_rdma_cleanup_event(struct qede_dev *edev) |
cee9fbd8 | 75 | { |
bbfcd1e8 MK |
76 | struct list_head *head = &edev->rdma_info.rdma_event_list; |
77 | struct qede_rdma_event_work *event_node; | |
cee9fbd8 | 78 | |
bbfcd1e8 | 79 | flush_workqueue(edev->rdma_info.rdma_wq); |
cee9fbd8 | 80 | while (!list_empty(head)) { |
bbfcd1e8 | 81 | event_node = list_entry(head->next, struct qede_rdma_event_work, |
cee9fbd8 RA |
82 | list); |
83 | cancel_work_sync(&event_node->work); | |
84 | list_del(&event_node->list); | |
85 | kfree(event_node); | |
86 | } | |
87 | } | |
88 | ||
af6565ad MK |
89 | static void qede_rdma_complete_event(struct kref *ref) |
90 | { | |
91 | struct qede_rdma_dev *rdma_dev = | |
92 | container_of(ref, struct qede_rdma_dev, refcnt); | |
93 | ||
94 | /* no more events will be added after this */ | |
95 | complete(&rdma_dev->event_comp); | |
96 | } | |
97 | ||
bbfcd1e8 | 98 | static void qede_rdma_destroy_wq(struct qede_dev *edev) |
cee9fbd8 | 99 | { |
af6565ad MK |
100 | /* Avoid race with add_event flow, make sure it finishes before |
101 | * we start accessing the list and cleaning up the work | |
102 | */ | |
103 | kref_put(&edev->rdma_info.refcnt, qede_rdma_complete_event); | |
104 | wait_for_completion(&edev->rdma_info.event_comp); | |
105 | ||
bbfcd1e8 MK |
106 | qede_rdma_cleanup_event(edev); |
107 | destroy_workqueue(edev->rdma_info.rdma_wq); | |
cee9fbd8 RA |
108 | } |
109 | ||
ccc67ef5 | 110 | int qede_rdma_dev_add(struct qede_dev *edev, bool recovery) |
cee9fbd8 | 111 | { |
ccc67ef5 | 112 | int rc; |
cee9fbd8 | 113 | |
ccc67ef5 TT |
114 | if (!qede_rdma_supported(edev)) |
115 | return 0; | |
cee9fbd8 | 116 | |
ccc67ef5 TT |
117 | /* Cannot start qedr while recovering since it wasn't fully stopped */ |
118 | if (recovery) | |
119 | return 0; | |
120 | ||
121 | rc = qede_rdma_create_wq(edev); | |
122 | if (rc) | |
123 | return rc; | |
124 | ||
125 | INIT_LIST_HEAD(&edev->rdma_info.entry); | |
126 | mutex_lock(&qedr_dev_list_lock); | |
127 | list_add_tail(&edev->rdma_info.entry, &qedr_dev_list); | |
128 | _qede_rdma_dev_add(edev); | |
129 | mutex_unlock(&qedr_dev_list_lock); | |
cee9fbd8 RA |
130 | |
131 | return rc; | |
132 | } | |
133 | ||
bbfcd1e8 | 134 | static void _qede_rdma_dev_remove(struct qede_dev *edev) |
cee9fbd8 RA |
135 | { |
136 | if (qedr_drv && qedr_drv->remove && edev->rdma_info.qedr_dev) | |
137 | qedr_drv->remove(edev->rdma_info.qedr_dev); | |
cee9fbd8 RA |
138 | } |
139 | ||
ccc67ef5 | 140 | void qede_rdma_dev_remove(struct qede_dev *edev, bool recovery) |
cee9fbd8 | 141 | { |
bbfcd1e8 | 142 | if (!qede_rdma_supported(edev)) |
cee9fbd8 RA |
143 | return; |
144 | ||
ccc67ef5 TT |
145 | /* Cannot remove qedr while recovering since it wasn't fully stopped */ |
146 | if (!recovery) { | |
147 | qede_rdma_destroy_wq(edev); | |
148 | mutex_lock(&qedr_dev_list_lock); | |
149 | if (!edev->rdma_info.exp_recovery) | |
150 | _qede_rdma_dev_remove(edev); | |
151 | edev->rdma_info.qedr_dev = NULL; | |
152 | list_del(&edev->rdma_info.entry); | |
153 | mutex_unlock(&qedr_dev_list_lock); | |
154 | } else { | |
155 | if (!edev->rdma_info.exp_recovery) { | |
156 | mutex_lock(&qedr_dev_list_lock); | |
157 | _qede_rdma_dev_remove(edev); | |
158 | mutex_unlock(&qedr_dev_list_lock); | |
159 | } | |
160 | edev->rdma_info.exp_recovery = true; | |
161 | } | |
cee9fbd8 RA |
162 | } |
163 | ||
bbfcd1e8 | 164 | static void _qede_rdma_dev_open(struct qede_dev *edev) |
cee9fbd8 RA |
165 | { |
166 | if (qedr_drv && edev->rdma_info.qedr_dev && qedr_drv->notify) | |
167 | qedr_drv->notify(edev->rdma_info.qedr_dev, QEDE_UP); | |
168 | } | |
169 | ||
bbfcd1e8 | 170 | static void qede_rdma_dev_open(struct qede_dev *edev) |
cee9fbd8 | 171 | { |
bbfcd1e8 | 172 | if (!qede_rdma_supported(edev)) |
cee9fbd8 RA |
173 | return; |
174 | ||
175 | mutex_lock(&qedr_dev_list_lock); | |
bbfcd1e8 | 176 | _qede_rdma_dev_open(edev); |
cee9fbd8 RA |
177 | mutex_unlock(&qedr_dev_list_lock); |
178 | } | |
179 | ||
bbfcd1e8 | 180 | static void _qede_rdma_dev_close(struct qede_dev *edev) |
cee9fbd8 RA |
181 | { |
182 | if (qedr_drv && edev->rdma_info.qedr_dev && qedr_drv->notify) | |
183 | qedr_drv->notify(edev->rdma_info.qedr_dev, QEDE_DOWN); | |
184 | } | |
185 | ||
bbfcd1e8 | 186 | static void qede_rdma_dev_close(struct qede_dev *edev) |
cee9fbd8 | 187 | { |
bbfcd1e8 | 188 | if (!qede_rdma_supported(edev)) |
cee9fbd8 RA |
189 | return; |
190 | ||
191 | mutex_lock(&qedr_dev_list_lock); | |
bbfcd1e8 | 192 | _qede_rdma_dev_close(edev); |
cee9fbd8 RA |
193 | mutex_unlock(&qedr_dev_list_lock); |
194 | } | |
195 | ||
bbfcd1e8 | 196 | static void qede_rdma_dev_shutdown(struct qede_dev *edev) |
cee9fbd8 | 197 | { |
bbfcd1e8 | 198 | if (!qede_rdma_supported(edev)) |
cee9fbd8 RA |
199 | return; |
200 | ||
201 | mutex_lock(&qedr_dev_list_lock); | |
202 | if (qedr_drv && edev->rdma_info.qedr_dev && qedr_drv->notify) | |
203 | qedr_drv->notify(edev->rdma_info.qedr_dev, QEDE_CLOSE); | |
204 | mutex_unlock(&qedr_dev_list_lock); | |
205 | } | |
206 | ||
bbfcd1e8 | 207 | int qede_rdma_register_driver(struct qedr_driver *drv) |
cee9fbd8 RA |
208 | { |
209 | struct qede_dev *edev; | |
210 | u8 qedr_counter = 0; | |
211 | ||
212 | mutex_lock(&qedr_dev_list_lock); | |
213 | if (qedr_drv) { | |
214 | mutex_unlock(&qedr_dev_list_lock); | |
215 | return -EINVAL; | |
216 | } | |
217 | qedr_drv = drv; | |
218 | ||
219 | list_for_each_entry(edev, &qedr_dev_list, rdma_info.entry) { | |
220 | struct net_device *ndev; | |
221 | ||
222 | qedr_counter++; | |
bbfcd1e8 | 223 | _qede_rdma_dev_add(edev); |
cee9fbd8 RA |
224 | ndev = edev->ndev; |
225 | if (netif_running(ndev) && netif_oper_up(ndev)) | |
bbfcd1e8 | 226 | _qede_rdma_dev_open(edev); |
cee9fbd8 RA |
227 | } |
228 | mutex_unlock(&qedr_dev_list_lock); | |
229 | ||
bbfcd1e8 | 230 | pr_notice("qedr: discovered and registered %d RDMA funcs\n", |
22b1ae61 | 231 | qedr_counter); |
cee9fbd8 RA |
232 | |
233 | return 0; | |
234 | } | |
bbfcd1e8 | 235 | EXPORT_SYMBOL(qede_rdma_register_driver); |
cee9fbd8 | 236 | |
bbfcd1e8 | 237 | void qede_rdma_unregister_driver(struct qedr_driver *drv) |
cee9fbd8 RA |
238 | { |
239 | struct qede_dev *edev; | |
240 | ||
241 | mutex_lock(&qedr_dev_list_lock); | |
242 | list_for_each_entry(edev, &qedr_dev_list, rdma_info.entry) { | |
ccc67ef5 TT |
243 | /* If device has experienced recovery it was already removed */ |
244 | if (edev->rdma_info.qedr_dev && !edev->rdma_info.exp_recovery) | |
bbfcd1e8 | 245 | _qede_rdma_dev_remove(edev); |
cee9fbd8 RA |
246 | } |
247 | qedr_drv = NULL; | |
248 | mutex_unlock(&qedr_dev_list_lock); | |
249 | } | |
bbfcd1e8 | 250 | EXPORT_SYMBOL(qede_rdma_unregister_driver); |
cee9fbd8 | 251 | |
bbfcd1e8 | 252 | static void qede_rdma_changeaddr(struct qede_dev *edev) |
cee9fbd8 | 253 | { |
bbfcd1e8 | 254 | if (!qede_rdma_supported(edev)) |
cee9fbd8 RA |
255 | return; |
256 | ||
257 | if (qedr_drv && edev->rdma_info.qedr_dev && qedr_drv->notify) | |
258 | qedr_drv->notify(edev->rdma_info.qedr_dev, QEDE_CHANGE_ADDR); | |
259 | } | |
260 | ||
bbfcd1e8 MK |
261 | static struct qede_rdma_event_work * |
262 | qede_rdma_get_free_event_node(struct qede_dev *edev) | |
cee9fbd8 | 263 | { |
bbfcd1e8 | 264 | struct qede_rdma_event_work *event_node = NULL; |
cee9fbd8 RA |
265 | struct list_head *list_node = NULL; |
266 | bool found = false; | |
267 | ||
bbfcd1e8 MK |
268 | list_for_each(list_node, &edev->rdma_info.rdma_event_list) { |
269 | event_node = list_entry(list_node, struct qede_rdma_event_work, | |
cee9fbd8 RA |
270 | list); |
271 | if (!work_pending(&event_node->work)) { | |
272 | found = true; | |
273 | break; | |
274 | } | |
275 | } | |
276 | ||
277 | if (!found) { | |
090477e4 | 278 | event_node = kzalloc(sizeof(*event_node), GFP_ATOMIC); |
cee9fbd8 RA |
279 | if (!event_node) { |
280 | DP_NOTICE(edev, | |
bbfcd1e8 | 281 | "qedr: Could not allocate memory for rdma work\n"); |
cee9fbd8 RA |
282 | return NULL; |
283 | } | |
284 | list_add_tail(&event_node->list, | |
bbfcd1e8 | 285 | &edev->rdma_info.rdma_event_list); |
cee9fbd8 RA |
286 | } |
287 | ||
288 | return event_node; | |
289 | } | |
290 | ||
bbfcd1e8 | 291 | static void qede_rdma_handle_event(struct work_struct *work) |
cee9fbd8 | 292 | { |
bbfcd1e8 MK |
293 | struct qede_rdma_event_work *event_node; |
294 | enum qede_rdma_event event; | |
cee9fbd8 RA |
295 | struct qede_dev *edev; |
296 | ||
bbfcd1e8 | 297 | event_node = container_of(work, struct qede_rdma_event_work, work); |
cee9fbd8 RA |
298 | event = event_node->event; |
299 | edev = event_node->ptr; | |
300 | ||
301 | switch (event) { | |
302 | case QEDE_UP: | |
bbfcd1e8 | 303 | qede_rdma_dev_open(edev); |
cee9fbd8 RA |
304 | break; |
305 | case QEDE_DOWN: | |
bbfcd1e8 | 306 | qede_rdma_dev_close(edev); |
cee9fbd8 RA |
307 | break; |
308 | case QEDE_CLOSE: | |
bbfcd1e8 | 309 | qede_rdma_dev_shutdown(edev); |
cee9fbd8 RA |
310 | break; |
311 | case QEDE_CHANGE_ADDR: | |
bbfcd1e8 | 312 | qede_rdma_changeaddr(edev); |
cee9fbd8 RA |
313 | break; |
314 | default: | |
bbfcd1e8 | 315 | DP_NOTICE(edev, "Invalid rdma event %d", event); |
cee9fbd8 RA |
316 | } |
317 | } | |
318 | ||
bbfcd1e8 MK |
319 | static void qede_rdma_add_event(struct qede_dev *edev, |
320 | enum qede_rdma_event event) | |
cee9fbd8 | 321 | { |
bbfcd1e8 | 322 | struct qede_rdma_event_work *event_node; |
cee9fbd8 | 323 | |
ccc67ef5 TT |
324 | /* If a recovery was experienced avoid adding the event */ |
325 | if (edev->rdma_info.exp_recovery) | |
326 | return; | |
327 | ||
cee9fbd8 RA |
328 | if (!edev->rdma_info.qedr_dev) |
329 | return; | |
330 | ||
af6565ad MK |
331 | /* We don't want the cleanup flow to start while we're allocating and |
332 | * scheduling the work | |
333 | */ | |
334 | if (!kref_get_unless_zero(&edev->rdma_info.refcnt)) | |
335 | return; /* already being destroyed */ | |
336 | ||
bbfcd1e8 | 337 | event_node = qede_rdma_get_free_event_node(edev); |
cee9fbd8 | 338 | if (!event_node) |
af6565ad | 339 | goto out; |
cee9fbd8 RA |
340 | |
341 | event_node->event = event; | |
342 | event_node->ptr = edev; | |
343 | ||
bbfcd1e8 MK |
344 | INIT_WORK(&event_node->work, qede_rdma_handle_event); |
345 | queue_work(edev->rdma_info.rdma_wq, &event_node->work); | |
af6565ad MK |
346 | |
347 | out: | |
348 | kref_put(&edev->rdma_info.refcnt, qede_rdma_complete_event); | |
cee9fbd8 RA |
349 | } |
350 | ||
bbfcd1e8 | 351 | void qede_rdma_dev_event_open(struct qede_dev *edev) |
cee9fbd8 | 352 | { |
bbfcd1e8 | 353 | qede_rdma_add_event(edev, QEDE_UP); |
cee9fbd8 RA |
354 | } |
355 | ||
bbfcd1e8 | 356 | void qede_rdma_dev_event_close(struct qede_dev *edev) |
cee9fbd8 | 357 | { |
bbfcd1e8 | 358 | qede_rdma_add_event(edev, QEDE_DOWN); |
cee9fbd8 RA |
359 | } |
360 | ||
bbfcd1e8 | 361 | void qede_rdma_event_changeaddr(struct qede_dev *edev) |
cee9fbd8 | 362 | { |
bbfcd1e8 | 363 | qede_rdma_add_event(edev, QEDE_CHANGE_ADDR); |
cee9fbd8 | 364 | } |