net: hns3: hclge_inform_reset_assert_to_vf() can be static
[linux-2.6-block.git] / drivers / net / ethernet / hisilicon / hns3 / hns3pf / hclge_mbx.c
1 // SPDX-License-Identifier: GPL-2.0+
2 // Copyright (c) 2016-2017 Hisilicon Limited.
3
4 #include "hclge_main.h"
5 #include "hclge_mbx.h"
6 #include "hnae3.h"
7
8 /* hclge_gen_resp_to_vf: used to generate a synchronous response to VF when PF
9  * receives a mailbox message from VF.
10  * @vport: pointer to struct hclge_vport
11  * @vf_to_pf_req: pointer to hclge_mbx_vf_to_pf_cmd of the original mailbox
12  *                message
13  * @resp_status: indicate to VF whether its request success(0) or failed.
14  */
15 static int hclge_gen_resp_to_vf(struct hclge_vport *vport,
16                                 struct hclge_mbx_vf_to_pf_cmd *vf_to_pf_req,
17                                 int resp_status,
18                                 u8 *resp_data, u16 resp_data_len)
19 {
20         struct hclge_mbx_pf_to_vf_cmd *resp_pf_to_vf;
21         struct hclge_dev *hdev = vport->back;
22         enum hclge_cmd_status status;
23         struct hclge_desc desc;
24
25         resp_pf_to_vf = (struct hclge_mbx_pf_to_vf_cmd *)desc.data;
26
27         if (resp_data_len > HCLGE_MBX_MAX_RESP_DATA_SIZE) {
28                 dev_err(&hdev->pdev->dev,
29                         "PF fail to gen resp to VF len %d exceeds max len %d\n",
30                         resp_data_len,
31                         HCLGE_MBX_MAX_RESP_DATA_SIZE);
32         }
33
34         hclge_cmd_setup_basic_desc(&desc, HCLGEVF_OPC_MBX_PF_TO_VF, false);
35
36         resp_pf_to_vf->dest_vfid = vf_to_pf_req->mbx_src_vfid;
37         resp_pf_to_vf->msg_len = vf_to_pf_req->msg_len;
38
39         resp_pf_to_vf->msg[0] = HCLGE_MBX_PF_VF_RESP;
40         resp_pf_to_vf->msg[1] = vf_to_pf_req->msg[0];
41         resp_pf_to_vf->msg[2] = vf_to_pf_req->msg[1];
42         resp_pf_to_vf->msg[3] = (resp_status == 0) ? 0 : 1;
43
44         if (resp_data && resp_data_len > 0)
45                 memcpy(&resp_pf_to_vf->msg[4], resp_data, resp_data_len);
46
47         status = hclge_cmd_send(&hdev->hw, &desc, 1);
48         if (status)
49                 dev_err(&hdev->pdev->dev,
50                         "PF failed(=%d) to send response to VF\n", status);
51
52         return status;
53 }
54
55 static int hclge_send_mbx_msg(struct hclge_vport *vport, u8 *msg, u16 msg_len,
56                               u16 mbx_opcode, u8 dest_vfid)
57 {
58         struct hclge_mbx_pf_to_vf_cmd *resp_pf_to_vf;
59         struct hclge_dev *hdev = vport->back;
60         enum hclge_cmd_status status;
61         struct hclge_desc desc;
62
63         resp_pf_to_vf = (struct hclge_mbx_pf_to_vf_cmd *)desc.data;
64
65         hclge_cmd_setup_basic_desc(&desc, HCLGEVF_OPC_MBX_PF_TO_VF, false);
66
67         resp_pf_to_vf->dest_vfid = dest_vfid;
68         resp_pf_to_vf->msg_len = msg_len;
69         resp_pf_to_vf->msg[0] = mbx_opcode;
70
71         memcpy(&resp_pf_to_vf->msg[1], msg, msg_len);
72
73         status = hclge_cmd_send(&hdev->hw, &desc, 1);
74         if (status)
75                 dev_err(&hdev->pdev->dev,
76                         "PF failed(=%d) to send mailbox message to VF\n",
77                         status);
78
79         return status;
80 }
81
82 static int hclge_inform_reset_assert_to_vf(struct hclge_vport *vport)
83 {
84         u8 msg_data[2];
85         u8 dest_vfid;
86
87         dest_vfid = (u8)vport->vport_id;
88
89         /* send this requested info to VF */
90         return hclge_send_mbx_msg(vport, msg_data, sizeof(u8),
91                                   HCLGE_MBX_ASSERTING_RESET, dest_vfid);
92 }
93
94 static void hclge_free_vector_ring_chain(struct hnae3_ring_chain_node *head)
95 {
96         struct hnae3_ring_chain_node *chain_tmp, *chain;
97
98         chain = head->next;
99
100         while (chain) {
101                 chain_tmp = chain->next;
102                 kzfree(chain);
103                 chain = chain_tmp;
104         }
105 }
106
107 /* hclge_get_ring_chain_from_mbx: get ring type & tqpid from mailbox message
108  * msg[0]: opcode
109  * msg[1]: <not relevant to this function>
110  * msg[2]: ring_num
111  * msg[3]: first ring type (TX|RX)
112  * msg[4]: first tqp id
113  * msg[5] ~ msg[14]: other ring type and tqp id
114  */
115 static int hclge_get_ring_chain_from_mbx(
116                         struct hclge_mbx_vf_to_pf_cmd *req,
117                         struct hnae3_ring_chain_node *ring_chain,
118                         struct hclge_vport *vport)
119 {
120         struct hnae3_ring_chain_node *cur_chain, *new_chain;
121         int ring_num;
122         int i;
123
124         ring_num = req->msg[2];
125
126         if (ring_num > ((HCLGE_MBX_VF_MSG_DATA_NUM -
127                 HCLGE_MBX_RING_MAP_BASIC_MSG_NUM) /
128                 HCLGE_MBX_RING_NODE_VARIABLE_NUM))
129                 return -ENOMEM;
130
131         hnae_set_bit(ring_chain->flag, HNAE3_RING_TYPE_B, req->msg[3]);
132         ring_chain->tqp_index =
133                         hclge_get_queue_id(vport->nic.kinfo.tqp[req->msg[4]]);
134         hnae_set_field(ring_chain->int_gl_idx, HCLGE_INT_GL_IDX_M,
135                        HCLGE_INT_GL_IDX_S,
136                        req->msg[5]);
137
138         cur_chain = ring_chain;
139
140         for (i = 1; i < ring_num; i++) {
141                 new_chain = kzalloc(sizeof(*new_chain), GFP_KERNEL);
142                 if (!new_chain)
143                         goto err;
144
145                 hnae_set_bit(new_chain->flag, HNAE3_RING_TYPE_B,
146                              req->msg[HCLGE_MBX_RING_NODE_VARIABLE_NUM * i +
147                              HCLGE_MBX_RING_MAP_BASIC_MSG_NUM]);
148
149                 new_chain->tqp_index =
150                 hclge_get_queue_id(vport->nic.kinfo.tqp
151                         [req->msg[HCLGE_MBX_RING_NODE_VARIABLE_NUM * i +
152                         HCLGE_MBX_RING_MAP_BASIC_MSG_NUM + 1]]);
153
154                 hnae_set_field(new_chain->int_gl_idx, HCLGE_INT_GL_IDX_M,
155                                HCLGE_INT_GL_IDX_S,
156                                req->msg[HCLGE_MBX_RING_NODE_VARIABLE_NUM * i +
157                                HCLGE_MBX_RING_MAP_BASIC_MSG_NUM + 2]);
158
159                 cur_chain->next = new_chain;
160                 cur_chain = new_chain;
161         }
162
163         return 0;
164 err:
165         hclge_free_vector_ring_chain(ring_chain);
166         return -ENOMEM;
167 }
168
169 static int hclge_map_unmap_ring_to_vf_vector(struct hclge_vport *vport, bool en,
170                                              struct hclge_mbx_vf_to_pf_cmd *req)
171 {
172         struct hnae3_ring_chain_node ring_chain;
173         int vector_id = req->msg[1];
174         int ret;
175
176         memset(&ring_chain, 0, sizeof(ring_chain));
177         ret = hclge_get_ring_chain_from_mbx(req, &ring_chain, vport);
178         if (ret)
179                 return ret;
180
181         ret = hclge_bind_ring_with_vector(vport, vector_id, en, &ring_chain);
182         if (ret)
183                 return ret;
184
185         hclge_free_vector_ring_chain(&ring_chain);
186
187         return 0;
188 }
189
190 static int hclge_set_vf_promisc_mode(struct hclge_vport *vport,
191                                      struct hclge_mbx_vf_to_pf_cmd *req)
192 {
193         bool en = req->msg[1] ? true : false;
194         struct hclge_promisc_param param;
195
196         /* always enable broadcast promisc bit */
197         hclge_promisc_param_init(&param, en, en, true, vport->vport_id);
198         return hclge_cmd_set_promisc_mode(vport->back, &param);
199 }
200
201 static int hclge_set_vf_uc_mac_addr(struct hclge_vport *vport,
202                                     struct hclge_mbx_vf_to_pf_cmd *mbx_req,
203                                     bool gen_resp)
204 {
205         const u8 *mac_addr = (const u8 *)(&mbx_req->msg[2]);
206         struct hclge_dev *hdev = vport->back;
207         int status;
208
209         if (mbx_req->msg[1] == HCLGE_MBX_MAC_VLAN_UC_MODIFY) {
210                 const u8 *old_addr = (const u8 *)(&mbx_req->msg[8]);
211
212                 hclge_rm_uc_addr_common(vport, old_addr);
213                 status = hclge_add_uc_addr_common(vport, mac_addr);
214                 if (status)
215                         hclge_add_uc_addr_common(vport, old_addr);
216         } else if (mbx_req->msg[1] == HCLGE_MBX_MAC_VLAN_UC_ADD) {
217                 status = hclge_add_uc_addr_common(vport, mac_addr);
218         } else if (mbx_req->msg[1] == HCLGE_MBX_MAC_VLAN_UC_REMOVE) {
219                 status = hclge_rm_uc_addr_common(vport, mac_addr);
220         } else {
221                 dev_err(&hdev->pdev->dev,
222                         "failed to set unicast mac addr, unknown subcode %d\n",
223                         mbx_req->msg[1]);
224                 return -EIO;
225         }
226
227         if (gen_resp)
228                 hclge_gen_resp_to_vf(vport, mbx_req, status, NULL, 0);
229
230         return 0;
231 }
232
233 static int hclge_set_vf_mc_mac_addr(struct hclge_vport *vport,
234                                     struct hclge_mbx_vf_to_pf_cmd *mbx_req,
235                                     bool gen_resp)
236 {
237         const u8 *mac_addr = (const u8 *)(&mbx_req->msg[2]);
238         struct hclge_dev *hdev = vport->back;
239         int status;
240
241         if (mbx_req->msg[1] == HCLGE_MBX_MAC_VLAN_MC_ADD) {
242                 status = hclge_add_mc_addr_common(vport, mac_addr);
243         } else if (mbx_req->msg[1] == HCLGE_MBX_MAC_VLAN_MC_REMOVE) {
244                 status = hclge_rm_mc_addr_common(vport, mac_addr);
245         } else if (mbx_req->msg[1] == HCLGE_MBX_MAC_VLAN_MC_FUNC_MTA_ENABLE) {
246                 u8 func_id = vport->vport_id;
247                 bool enable = mbx_req->msg[2];
248
249                 status = hclge_cfg_func_mta_filter(hdev, func_id, enable);
250         } else {
251                 dev_err(&hdev->pdev->dev,
252                         "failed to set mcast mac addr, unknown subcode %d\n",
253                         mbx_req->msg[1]);
254                 return -EIO;
255         }
256
257         if (gen_resp)
258                 hclge_gen_resp_to_vf(vport, mbx_req, status, NULL, 0);
259
260         return 0;
261 }
262
263 static int hclge_set_vf_vlan_cfg(struct hclge_vport *vport,
264                                  struct hclge_mbx_vf_to_pf_cmd *mbx_req,
265                                  bool gen_resp)
266 {
267         struct hclge_dev *hdev = vport->back;
268         int status = 0;
269
270         if (mbx_req->msg[1] == HCLGE_MBX_VLAN_FILTER) {
271                 u16 vlan, proto;
272                 bool is_kill;
273
274                 is_kill = !!mbx_req->msg[2];
275                 memcpy(&vlan, &mbx_req->msg[3], sizeof(vlan));
276                 memcpy(&proto, &mbx_req->msg[5], sizeof(proto));
277                 status = hclge_set_vf_vlan_common(hdev, vport->vport_id,
278                                                   is_kill, vlan, 0,
279                                                   cpu_to_be16(proto));
280         }
281
282         if (gen_resp)
283                 status = hclge_gen_resp_to_vf(vport, mbx_req, status, NULL, 0);
284
285         return status;
286 }
287
288 static int hclge_get_vf_tcinfo(struct hclge_vport *vport,
289                                struct hclge_mbx_vf_to_pf_cmd *mbx_req,
290                                bool gen_resp)
291 {
292         struct hclge_dev *hdev = vport->back;
293         int ret;
294
295         ret = hclge_gen_resp_to_vf(vport, mbx_req, 0, &hdev->hw_tc_map,
296                                    sizeof(u8));
297
298         return ret;
299 }
300
301 static int hclge_get_vf_queue_info(struct hclge_vport *vport,
302                                    struct hclge_mbx_vf_to_pf_cmd *mbx_req,
303                                    bool gen_resp)
304 {
305 #define HCLGE_TQPS_RSS_INFO_LEN         8
306         u8 resp_data[HCLGE_TQPS_RSS_INFO_LEN];
307         struct hclge_dev *hdev = vport->back;
308
309         /* get the queue related info */
310         memcpy(&resp_data[0], &vport->alloc_tqps, sizeof(u16));
311         memcpy(&resp_data[2], &vport->nic.kinfo.rss_size, sizeof(u16));
312         memcpy(&resp_data[4], &hdev->num_desc, sizeof(u16));
313         memcpy(&resp_data[6], &hdev->rx_buf_len, sizeof(u16));
314
315         return hclge_gen_resp_to_vf(vport, mbx_req, 0, resp_data,
316                                     HCLGE_TQPS_RSS_INFO_LEN);
317 }
318
319 static int hclge_get_link_info(struct hclge_vport *vport,
320                                struct hclge_mbx_vf_to_pf_cmd *mbx_req)
321 {
322         struct hclge_dev *hdev = vport->back;
323         u16 link_status;
324         u8 msg_data[8];
325         u8 dest_vfid;
326         u16 duplex;
327
328         /* mac.link can only be 0 or 1 */
329         link_status = (u16)hdev->hw.mac.link;
330         duplex = hdev->hw.mac.duplex;
331         memcpy(&msg_data[0], &link_status, sizeof(u16));
332         memcpy(&msg_data[2], &hdev->hw.mac.speed, sizeof(u32));
333         memcpy(&msg_data[6], &duplex, sizeof(u16));
334         dest_vfid = mbx_req->mbx_src_vfid;
335
336         /* send this requested info to VF */
337         return hclge_send_mbx_msg(vport, msg_data, sizeof(msg_data),
338                                   HCLGE_MBX_LINK_STAT_CHANGE, dest_vfid);
339 }
340
341 static void hclge_mbx_reset_vf_queue(struct hclge_vport *vport,
342                                      struct hclge_mbx_vf_to_pf_cmd *mbx_req)
343 {
344         u16 queue_id;
345
346         memcpy(&queue_id, &mbx_req->msg[2], sizeof(queue_id));
347
348         hclge_reset_vf_queue(vport, queue_id);
349
350         /* send response msg to VF after queue reset complete*/
351         hclge_gen_resp_to_vf(vport, mbx_req, 0, NULL, 0);
352 }
353
354 static void hclge_reset_vf(struct hclge_vport *vport,
355                            struct hclge_mbx_vf_to_pf_cmd *mbx_req)
356 {
357         struct hclge_dev *hdev = vport->back;
358         int ret;
359
360         dev_warn(&hdev->pdev->dev, "PF received VF reset request from VF %d!",
361                  mbx_req->mbx_src_vfid);
362
363         /* Acknowledge VF that PF is now about to assert the reset for the VF.
364          * On receiving this message VF will get into pending state and will
365          * start polling for the hardware reset completion status.
366          */
367         ret = hclge_inform_reset_assert_to_vf(vport);
368         if (ret) {
369                 dev_err(&hdev->pdev->dev,
370                         "PF fail(%d) to inform VF(%d)of reset, reset failed!\n",
371                         ret, vport->vport_id);
372                 return;
373         }
374
375         dev_warn(&hdev->pdev->dev, "PF is now resetting VF %d.\n",
376                  mbx_req->mbx_src_vfid);
377         /* reset this virtual function */
378         hclge_func_reset_cmd(hdev, mbx_req->mbx_src_vfid);
379 }
380
381 void hclge_mbx_handler(struct hclge_dev *hdev)
382 {
383         struct hclge_cmq_ring *crq = &hdev->hw.cmq.crq;
384         struct hclge_mbx_vf_to_pf_cmd *req;
385         struct hclge_vport *vport;
386         struct hclge_desc *desc;
387         int ret, flag;
388
389         flag = le16_to_cpu(crq->desc[crq->next_to_use].flag);
390         /* handle all the mailbox requests in the queue */
391         while (hnae_get_bit(flag, HCLGE_CMDQ_RX_OUTVLD_B)) {
392                 desc = &crq->desc[crq->next_to_use];
393                 req = (struct hclge_mbx_vf_to_pf_cmd *)desc->data;
394
395                 vport = &hdev->vport[req->mbx_src_vfid];
396
397                 switch (req->msg[0]) {
398                 case HCLGE_MBX_MAP_RING_TO_VECTOR:
399                         ret = hclge_map_unmap_ring_to_vf_vector(vport, true,
400                                                                 req);
401                         break;
402                 case HCLGE_MBX_UNMAP_RING_TO_VECTOR:
403                         ret = hclge_map_unmap_ring_to_vf_vector(vport, false,
404                                                                 req);
405                         break;
406                 case HCLGE_MBX_SET_PROMISC_MODE:
407                         ret = hclge_set_vf_promisc_mode(vport, req);
408                         if (ret)
409                                 dev_err(&hdev->pdev->dev,
410                                         "PF fail(%d) to set VF promisc mode\n",
411                                         ret);
412                         break;
413                 case HCLGE_MBX_SET_UNICAST:
414                         ret = hclge_set_vf_uc_mac_addr(vport, req, true);
415                         if (ret)
416                                 dev_err(&hdev->pdev->dev,
417                                         "PF fail(%d) to set VF UC MAC Addr\n",
418                                         ret);
419                         break;
420                 case HCLGE_MBX_SET_MULTICAST:
421                         ret = hclge_set_vf_mc_mac_addr(vport, req, false);
422                         if (ret)
423                                 dev_err(&hdev->pdev->dev,
424                                         "PF fail(%d) to set VF MC MAC Addr\n",
425                                         ret);
426                         break;
427                 case HCLGE_MBX_SET_VLAN:
428                         ret = hclge_set_vf_vlan_cfg(vport, req, false);
429                         if (ret)
430                                 dev_err(&hdev->pdev->dev,
431                                         "PF failed(%d) to config VF's VLAN\n",
432                                         ret);
433                         break;
434                 case HCLGE_MBX_GET_QINFO:
435                         ret = hclge_get_vf_queue_info(vport, req, true);
436                         if (ret)
437                                 dev_err(&hdev->pdev->dev,
438                                         "PF failed(%d) to get Q info for VF\n",
439                                         ret);
440                         break;
441                 case HCLGE_MBX_GET_TCINFO:
442                         ret = hclge_get_vf_tcinfo(vport, req, true);
443                         if (ret)
444                                 dev_err(&hdev->pdev->dev,
445                                         "PF failed(%d) to get TC info for VF\n",
446                                         ret);
447                         break;
448                 case HCLGE_MBX_GET_LINK_STATUS:
449                         ret = hclge_get_link_info(vport, req);
450                         if (ret)
451                                 dev_err(&hdev->pdev->dev,
452                                         "PF fail(%d) to get link stat for VF\n",
453                                         ret);
454                         break;
455                 case HCLGE_MBX_QUEUE_RESET:
456                         hclge_mbx_reset_vf_queue(vport, req);
457                         break;
458                 case HCLGE_MBX_RESET:
459                         hclge_reset_vf(vport, req);
460                         break;
461                 default:
462                         dev_err(&hdev->pdev->dev,
463                                 "un-supported mailbox message, code = %d\n",
464                                 req->msg[0]);
465                         break;
466                 }
467                 crq->desc[crq->next_to_use].flag = 0;
468                 hclge_mbx_ring_ptr_move_crq(crq);
469                 flag = le16_to_cpu(crq->desc[crq->next_to_use].flag);
470         }
471
472         /* Write back CMDQ_RQ header pointer, M7 need this pointer */
473         hclge_write_dev(&hdev->hw, HCLGE_NIC_CRQ_HEAD_REG, crq->next_to_use);
474 }