Commit | Line | Data |
---|---|---|
e85b24b5 OG |
1 | /* |
2 | * Copyright (c) 2004, 2005, 2006 Voltaire, Inc. All rights reserved. | |
3ee07d27 | 3 | * Copyright (c) 2013-2014 Mellanox Technologies. All rights reserved. |
e85b24b5 OG |
4 | * |
5 | * This software is available to you under a choice of one of two | |
6 | * licenses. You may choose to be licensed under the terms of the GNU | |
7 | * General Public License (GPL) Version 2, available from the file | |
8 | * COPYING in the main directory of this source tree, or the | |
9 | * OpenIB.org BSD license below: | |
10 | * | |
11 | * Redistribution and use in source and binary forms, with or | |
12 | * without modification, are permitted provided that the following | |
13 | * conditions are met: | |
14 | * | |
15 | * - Redistributions of source code must retain the above | |
16 | * copyright notice, this list of conditions and the following | |
17 | * disclaimer. | |
18 | * | |
19 | * - Redistributions in binary form must reproduce the above | |
20 | * copyright notice, this list of conditions and the following | |
21 | * disclaimer in the documentation and/or other materials | |
22 | * provided with the distribution. | |
23 | * | |
24 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | |
25 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | |
26 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND | |
27 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS | |
28 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN | |
29 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN | |
30 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | |
31 | * SOFTWARE. | |
e85b24b5 OG |
32 | */ |
33 | #include <linux/kernel.h> | |
34 | #include <linux/slab.h> | |
35 | #include <linux/mm.h> | |
e85b24b5 OG |
36 | #include <linux/scatterlist.h> |
37 | #include <linux/kfifo.h> | |
38 | #include <scsi/scsi_cmnd.h> | |
39 | #include <scsi/scsi_host.h> | |
40 | ||
41 | #include "iscsi_iser.h" | |
42 | ||
e85b24b5 | 43 | /* Register user buffer memory and initialize passive rdma |
d77e6535 SG |
44 | * dto descriptor. Data size is stored in |
45 | * task->data[ISER_DIR_IN].data_len, Protection size | |
46 | * os stored in task->prot[ISER_DIR_IN].data_len | |
e85b24b5 | 47 | */ |
d77e6535 | 48 | static int iser_prepare_read_cmd(struct iscsi_task *task) |
e85b24b5 OG |
49 | |
50 | { | |
2261ec3d | 51 | struct iscsi_iser_task *iser_task = task->dd_data; |
b130eded | 52 | struct iser_mem_reg *mem_reg; |
e85b24b5 | 53 | int err; |
d3cf81f9 | 54 | struct iser_ctrl *hdr = &iser_task->desc.iser_header; |
2261ec3d | 55 | struct iser_data_buf *buf_in = &iser_task->data[ISER_DIR_IN]; |
e85b24b5 | 56 | |
2261ec3d | 57 | err = iser_dma_map_task_data(iser_task, |
e85b24b5 OG |
58 | buf_in, |
59 | ISER_DIR_IN, | |
60 | DMA_FROM_DEVICE); | |
61 | if (err) | |
62 | return err; | |
63 | ||
177e31bd SG |
64 | if (scsi_prot_sg_count(iser_task->sc)) { |
65 | struct iser_data_buf *pbuf_in = &iser_task->prot[ISER_DIR_IN]; | |
66 | ||
67 | err = iser_dma_map_task_data(iser_task, | |
68 | pbuf_in, | |
69 | ISER_DIR_IN, | |
70 | DMA_FROM_DEVICE); | |
71 | if (err) | |
72 | return err; | |
73 | } | |
74 | ||
b5f04b00 | 75 | err = iser_reg_rdma_mem(iser_task, ISER_DIR_IN, false); |
e85b24b5 OG |
76 | if (err) { |
77 | iser_err("Failed to set up Data-IN RDMA\n"); | |
78 | return err; | |
79 | } | |
b130eded | 80 | mem_reg = &iser_task->rdma_reg[ISER_DIR_IN]; |
e85b24b5 OG |
81 | |
82 | hdr->flags |= ISER_RSV; | |
b130eded | 83 | hdr->read_stag = cpu_to_be32(mem_reg->rkey); |
90a6684c | 84 | hdr->read_va = cpu_to_be64(mem_reg->sge.addr); |
e85b24b5 OG |
85 | |
86 | iser_dbg("Cmd itt:%d READ tags RKEY:%#.4X VA:%#llX\n", | |
b130eded | 87 | task->itt, mem_reg->rkey, |
90a6684c | 88 | (unsigned long long)mem_reg->sge.addr); |
e85b24b5 OG |
89 | |
90 | return 0; | |
91 | } | |
92 | ||
93 | /* Register user buffer memory and initialize passive rdma | |
d77e6535 SG |
94 | * dto descriptor. Data size is stored in |
95 | * task->data[ISER_DIR_OUT].data_len, Protection size | |
96 | * is stored at task->prot[ISER_DIR_OUT].data_len | |
e85b24b5 OG |
97 | */ |
98 | static int | |
2261ec3d | 99 | iser_prepare_write_cmd(struct iscsi_task *task, |
e85b24b5 OG |
100 | unsigned int imm_sz, |
101 | unsigned int unsol_sz, | |
102 | unsigned int edtl) | |
103 | { | |
2261ec3d | 104 | struct iscsi_iser_task *iser_task = task->dd_data; |
b130eded | 105 | struct iser_mem_reg *mem_reg; |
e85b24b5 | 106 | int err; |
d3cf81f9 | 107 | struct iser_ctrl *hdr = &iser_task->desc.iser_header; |
2261ec3d | 108 | struct iser_data_buf *buf_out = &iser_task->data[ISER_DIR_OUT]; |
f19624aa | 109 | struct ib_sge *tx_dsg = &iser_task->desc.tx_sg[1]; |
e85b24b5 | 110 | |
2261ec3d | 111 | err = iser_dma_map_task_data(iser_task, |
e85b24b5 OG |
112 | buf_out, |
113 | ISER_DIR_OUT, | |
114 | DMA_TO_DEVICE); | |
115 | if (err) | |
116 | return err; | |
117 | ||
177e31bd SG |
118 | if (scsi_prot_sg_count(iser_task->sc)) { |
119 | struct iser_data_buf *pbuf_out = &iser_task->prot[ISER_DIR_OUT]; | |
120 | ||
121 | err = iser_dma_map_task_data(iser_task, | |
122 | pbuf_out, | |
123 | ISER_DIR_OUT, | |
124 | DMA_TO_DEVICE); | |
125 | if (err) | |
126 | return err; | |
127 | } | |
128 | ||
b5f04b00 JD |
129 | err = iser_reg_rdma_mem(iser_task, ISER_DIR_OUT, |
130 | buf_out->data_len == imm_sz); | |
e85b24b5 OG |
131 | if (err != 0) { |
132 | iser_err("Failed to register write cmd RDMA mem\n"); | |
133 | return err; | |
134 | } | |
135 | ||
b130eded | 136 | mem_reg = &iser_task->rdma_reg[ISER_DIR_OUT]; |
e85b24b5 OG |
137 | |
138 | if (unsol_sz < edtl) { | |
139 | hdr->flags |= ISER_WSV; | |
e6e52aec SG |
140 | if (buf_out->data_len > imm_sz) { |
141 | hdr->write_stag = cpu_to_be32(mem_reg->rkey); | |
142 | hdr->write_va = cpu_to_be64(mem_reg->sge.addr + unsol_sz); | |
143 | } | |
e85b24b5 | 144 | |
4cb24c55 | 145 | iser_dbg("Cmd itt:%d, WRITE tags, RKEY:%#.4X VA:%#llX + unsol:%d\n", |
b130eded | 146 | task->itt, mem_reg->rkey, |
90a6684c | 147 | (unsigned long long)mem_reg->sge.addr, unsol_sz); |
e85b24b5 OG |
148 | } |
149 | ||
150 | if (imm_sz > 0) { | |
151 | iser_dbg("Cmd itt:%d, WRITE, adding imm.data sz: %d\n", | |
2261ec3d | 152 | task->itt, imm_sz); |
90a6684c | 153 | tx_dsg->addr = mem_reg->sge.addr; |
f19624aa | 154 | tx_dsg->length = imm_sz; |
90a6684c | 155 | tx_dsg->lkey = mem_reg->sge.lkey; |
f19624aa | 156 | iser_task->desc.num_sge = 2; |
e85b24b5 OG |
157 | } |
158 | ||
159 | return 0; | |
160 | } | |
161 | ||
e85b24b5 | 162 | /* creates a new tx descriptor and adds header regd buffer */ |
5716af6e | 163 | static void iser_create_send_desc(struct iser_conn *iser_conn, |
f19624aa | 164 | struct iser_tx_desc *tx_desc) |
e85b24b5 | 165 | { |
a4ee3539 | 166 | struct iser_device *device = iser_conn->ib_conn.device; |
e85b24b5 | 167 | |
f19624aa OG |
168 | ib_dma_sync_single_for_cpu(device->ib_device, |
169 | tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); | |
e85b24b5 | 170 | |
d3cf81f9 | 171 | memset(&tx_desc->iser_header, 0, sizeof(struct iser_ctrl)); |
e85b24b5 | 172 | tx_desc->iser_header.flags = ISER_VER; |
f19624aa | 173 | tx_desc->num_sge = 1; |
e85b24b5 OG |
174 | } |
175 | ||
5716af6e | 176 | static void iser_free_login_buf(struct iser_conn *iser_conn) |
986db0d6 | 177 | { |
a4ee3539 | 178 | struct iser_device *device = iser_conn->ib_conn.device; |
0f512b34 | 179 | struct iser_login_desc *desc = &iser_conn->login_desc; |
a4ee3539 | 180 | |
0f512b34 | 181 | if (!desc->req) |
986db0d6 SP |
182 | return; |
183 | ||
0f512b34 SG |
184 | ib_dma_unmap_single(device->ib_device, desc->req_dma, |
185 | ISCSI_DEF_MAX_RECV_SEG_LEN, DMA_TO_DEVICE); | |
986db0d6 | 186 | |
0f512b34 SG |
187 | ib_dma_unmap_single(device->ib_device, desc->rsp_dma, |
188 | ISER_RX_LOGIN_SIZE, DMA_FROM_DEVICE); | |
986db0d6 | 189 | |
0f512b34 SG |
190 | kfree(desc->req); |
191 | kfree(desc->rsp); | |
986db0d6 SP |
192 | |
193 | /* make sure we never redo any unmapping */ | |
0f512b34 SG |
194 | desc->req = NULL; |
195 | desc->rsp = NULL; | |
986db0d6 SP |
196 | } |
197 | ||
5716af6e | 198 | static int iser_alloc_login_buf(struct iser_conn *iser_conn) |
986db0d6 | 199 | { |
a4ee3539 | 200 | struct iser_device *device = iser_conn->ib_conn.device; |
0f512b34 SG |
201 | struct iser_login_desc *desc = &iser_conn->login_desc; |
202 | ||
203 | desc->req = kmalloc(ISCSI_DEF_MAX_RECV_SEG_LEN, GFP_KERNEL); | |
204 | if (!desc->req) | |
205 | return -ENOMEM; | |
206 | ||
207 | desc->req_dma = ib_dma_map_single(device->ib_device, desc->req, | |
208 | ISCSI_DEF_MAX_RECV_SEG_LEN, | |
209 | DMA_TO_DEVICE); | |
210 | if (ib_dma_mapping_error(device->ib_device, | |
211 | desc->req_dma)) | |
212 | goto free_req; | |
213 | ||
214 | desc->rsp = kmalloc(ISER_RX_LOGIN_SIZE, GFP_KERNEL); | |
215 | if (!desc->rsp) | |
216 | goto unmap_req; | |
217 | ||
218 | desc->rsp_dma = ib_dma_map_single(device->ib_device, desc->rsp, | |
219 | ISER_RX_LOGIN_SIZE, | |
220 | DMA_FROM_DEVICE); | |
221 | if (ib_dma_mapping_error(device->ib_device, | |
222 | desc->rsp_dma)) | |
223 | goto free_rsp; | |
224 | ||
986db0d6 SP |
225 | return 0; |
226 | ||
0f512b34 SG |
227 | free_rsp: |
228 | kfree(desc->rsp); | |
229 | unmap_req: | |
230 | ib_dma_unmap_single(device->ib_device, desc->req_dma, | |
231 | ISCSI_DEF_MAX_RECV_SEG_LEN, | |
232 | DMA_TO_DEVICE); | |
233 | free_req: | |
234 | kfree(desc->req); | |
986db0d6 | 235 | |
986db0d6 SP |
236 | return -ENOMEM; |
237 | } | |
f19624aa | 238 | |
5716af6e SG |
239 | int iser_alloc_rx_descriptors(struct iser_conn *iser_conn, |
240 | struct iscsi_session *session) | |
bcc60c38 OG |
241 | { |
242 | int i, j; | |
243 | u64 dma_addr; | |
244 | struct iser_rx_desc *rx_desc; | |
245 | struct ib_sge *rx_sg; | |
a4ee3539 SG |
246 | struct ib_conn *ib_conn = &iser_conn->ib_conn; |
247 | struct iser_device *device = ib_conn->device; | |
bcc60c38 | 248 | |
5716af6e SG |
249 | iser_conn->qp_max_recv_dtos = session->cmds_max; |
250 | iser_conn->qp_max_recv_dtos_mask = session->cmds_max - 1; /* cmds_max is 2^N */ | |
251 | iser_conn->min_posted_rx = iser_conn->qp_max_recv_dtos >> 2; | |
b7f04513 | 252 | |
f8db651d | 253 | if (device->reg_ops->alloc_reg_res(ib_conn, session->scsi_cmds_max, |
434dda42 | 254 | iser_conn->pages_per_mr)) |
b4e155ff | 255 | goto create_rdma_reg_res_failed; |
986db0d6 | 256 | |
5716af6e | 257 | if (iser_alloc_login_buf(iser_conn)) |
986db0d6 SP |
258 | goto alloc_login_buf_fail; |
259 | ||
6aabfa76 | 260 | iser_conn->num_rx_descs = session->cmds_max; |
6da2ec56 KC |
261 | iser_conn->rx_descs = kmalloc_array(iser_conn->num_rx_descs, |
262 | sizeof(struct iser_rx_desc), | |
263 | GFP_KERNEL); | |
5716af6e | 264 | if (!iser_conn->rx_descs) |
bcc60c38 OG |
265 | goto rx_desc_alloc_fail; |
266 | ||
5716af6e | 267 | rx_desc = iser_conn->rx_descs; |
bcc60c38 | 268 | |
5716af6e | 269 | for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++) { |
bcc60c38 OG |
270 | dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, |
271 | ISER_RX_PAYLOAD_SIZE, DMA_FROM_DEVICE); | |
272 | if (ib_dma_mapping_error(device->ib_device, dma_addr)) | |
273 | goto rx_desc_dma_map_failed; | |
274 | ||
275 | rx_desc->dma_addr = dma_addr; | |
cfeb91b3 | 276 | rx_desc->cqe.done = iser_task_rsp; |
bcc60c38 | 277 | rx_sg = &rx_desc->rx_sg; |
cfeb91b3 | 278 | rx_sg->addr = rx_desc->dma_addr; |
bcc60c38 | 279 | rx_sg->length = ISER_RX_PAYLOAD_SIZE; |
cfeb91b3 | 280 | rx_sg->lkey = device->pd->local_dma_lkey; |
bcc60c38 OG |
281 | } |
282 | ||
5716af6e | 283 | iser_conn->rx_desc_head = 0; |
bcc60c38 OG |
284 | return 0; |
285 | ||
286 | rx_desc_dma_map_failed: | |
5716af6e | 287 | rx_desc = iser_conn->rx_descs; |
bcc60c38 OG |
288 | for (j = 0; j < i; j++, rx_desc++) |
289 | ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, | |
986db0d6 | 290 | ISER_RX_PAYLOAD_SIZE, DMA_FROM_DEVICE); |
5716af6e SG |
291 | kfree(iser_conn->rx_descs); |
292 | iser_conn->rx_descs = NULL; | |
bcc60c38 | 293 | rx_desc_alloc_fail: |
5716af6e | 294 | iser_free_login_buf(iser_conn); |
986db0d6 | 295 | alloc_login_buf_fail: |
48afbff6 | 296 | device->reg_ops->free_reg_res(ib_conn); |
b4e155ff | 297 | create_rdma_reg_res_failed: |
bcc60c38 OG |
298 | iser_err("failed allocating rx descriptors / data buffers\n"); |
299 | return -ENOMEM; | |
300 | } | |
301 | ||
5716af6e | 302 | void iser_free_rx_descriptors(struct iser_conn *iser_conn) |
bcc60c38 OG |
303 | { |
304 | int i; | |
305 | struct iser_rx_desc *rx_desc; | |
a4ee3539 SG |
306 | struct ib_conn *ib_conn = &iser_conn->ib_conn; |
307 | struct iser_device *device = ib_conn->device; | |
bcc60c38 | 308 | |
48afbff6 SG |
309 | if (device->reg_ops->free_reg_res) |
310 | device->reg_ops->free_reg_res(ib_conn); | |
bcc60c38 | 311 | |
5716af6e SG |
312 | rx_desc = iser_conn->rx_descs; |
313 | for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++) | |
bcc60c38 | 314 | ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, |
986db0d6 | 315 | ISER_RX_PAYLOAD_SIZE, DMA_FROM_DEVICE); |
5716af6e | 316 | kfree(iser_conn->rx_descs); |
986db0d6 | 317 | /* make sure we never redo any unmapping */ |
5716af6e | 318 | iser_conn->rx_descs = NULL; |
986db0d6 | 319 | |
5716af6e | 320 | iser_free_login_buf(iser_conn); |
bcc60c38 OG |
321 | } |
322 | ||
89e984e2 | 323 | static int iser_post_rx_bufs(struct iscsi_conn *conn, struct iscsi_hdr *req) |
e85b24b5 | 324 | { |
5716af6e | 325 | struct iser_conn *iser_conn = conn->dd_data; |
a4ee3539 | 326 | struct ib_conn *ib_conn = &iser_conn->ib_conn; |
6a06a4b8 | 327 | struct iscsi_session *session = conn->session; |
e85b24b5 | 328 | |
89e984e2 OG |
329 | iser_dbg("req op %x flags %x\n", req->opcode, req->flags); |
330 | /* check if this is the last login - going to full feature phase */ | |
331 | if ((req->flags & ISCSI_FULL_FEATURE_PHASE) != ISCSI_FULL_FEATURE_PHASE) | |
332 | return 0; | |
e85b24b5 | 333 | |
89e984e2 | 334 | /* |
ff3dd52d SG |
335 | * Check that there is one posted recv buffer |
336 | * (for the last login response). | |
89e984e2 | 337 | */ |
a4ee3539 | 338 | WARN_ON(ib_conn->post_recv_buf_count != 1); |
bcc60c38 | 339 | |
6a06a4b8 OG |
340 | if (session->discovery_sess) { |
341 | iser_info("Discovery session, re-using login RX buffer\n"); | |
342 | return 0; | |
343 | } else | |
344 | iser_info("Normal session, posting batch of RX %d buffers\n", | |
5716af6e | 345 | iser_conn->min_posted_rx); |
6a06a4b8 | 346 | |
e85b24b5 | 347 | /* Initial post receive buffers */ |
5716af6e | 348 | if (iser_post_recvm(iser_conn, iser_conn->min_posted_rx)) |
bcc60c38 OG |
349 | return -ENOMEM; |
350 | ||
e85b24b5 OG |
351 | return 0; |
352 | } | |
353 | ||
6ec9d4d2 | 354 | static inline bool iser_signal_comp(u8 sig_count) |
6df5a128 SG |
355 | { |
356 | return ((sig_count % ISER_SIGNAL_CMD_COUNT) == 0); | |
357 | } | |
358 | ||
e85b24b5 OG |
359 | /** |
360 | * iser_send_command - send command PDU | |
361 | */ | |
2747fdb2 | 362 | int iser_send_command(struct iscsi_conn *conn, |
2261ec3d | 363 | struct iscsi_task *task) |
e85b24b5 | 364 | { |
5716af6e | 365 | struct iser_conn *iser_conn = conn->dd_data; |
2261ec3d | 366 | struct iscsi_iser_task *iser_task = task->dd_data; |
e85b24b5 | 367 | unsigned long edtl; |
bcc60c38 | 368 | int err; |
177e31bd | 369 | struct iser_data_buf *data_buf, *prot_buf; |
12352183 | 370 | struct iscsi_scsi_req *hdr = (struct iscsi_scsi_req *)task->hdr; |
2261ec3d | 371 | struct scsi_cmnd *sc = task->sc; |
f19624aa | 372 | struct iser_tx_desc *tx_desc = &iser_task->desc; |
6ec9d4d2 | 373 | u8 sig_count = ++iser_conn->ib_conn.sig_count; |
e85b24b5 | 374 | |
e85b24b5 OG |
375 | edtl = ntohl(hdr->data_length); |
376 | ||
377 | /* build the tx desc regd header and add it to the tx desc dto */ | |
f19624aa | 378 | tx_desc->type = ISCSI_TX_SCSI_COMMAND; |
cfeb91b3 | 379 | tx_desc->cqe.done = iser_cmd_comp; |
5716af6e | 380 | iser_create_send_desc(iser_conn, tx_desc); |
e85b24b5 | 381 | |
177e31bd | 382 | if (hdr->flags & ISCSI_FLAG_CMD_READ) { |
2261ec3d | 383 | data_buf = &iser_task->data[ISER_DIR_IN]; |
177e31bd SG |
384 | prot_buf = &iser_task->prot[ISER_DIR_IN]; |
385 | } else { | |
2261ec3d | 386 | data_buf = &iser_task->data[ISER_DIR_OUT]; |
177e31bd SG |
387 | prot_buf = &iser_task->prot[ISER_DIR_OUT]; |
388 | } | |
e85b24b5 | 389 | |
da9c0c77 | 390 | if (scsi_sg_count(sc)) { /* using a scatter list */ |
e3784bd1 | 391 | data_buf->sg = scsi_sglist(sc); |
da9c0c77 | 392 | data_buf->size = scsi_sg_count(sc); |
e85b24b5 | 393 | } |
da9c0c77 | 394 | data_buf->data_len = scsi_bufflen(sc); |
e85b24b5 | 395 | |
177e31bd | 396 | if (scsi_prot_sg_count(sc)) { |
e3784bd1 | 397 | prot_buf->sg = scsi_prot_sglist(sc); |
177e31bd | 398 | prot_buf->size = scsi_prot_sg_count(sc); |
a065fe6a SG |
399 | prot_buf->data_len = (data_buf->data_len >> |
400 | ilog2(sc->device->sector_size)) * 8; | |
177e31bd SG |
401 | } |
402 | ||
e85b24b5 | 403 | if (hdr->flags & ISCSI_FLAG_CMD_READ) { |
d77e6535 | 404 | err = iser_prepare_read_cmd(task); |
e85b24b5 OG |
405 | if (err) |
406 | goto send_command_error; | |
407 | } | |
408 | if (hdr->flags & ISCSI_FLAG_CMD_WRITE) { | |
2261ec3d MC |
409 | err = iser_prepare_write_cmd(task, |
410 | task->imm_count, | |
411 | task->imm_count + | |
0f9c7449 | 412 | task->unsol_r2t.data_length, |
e85b24b5 OG |
413 | edtl); |
414 | if (err) | |
415 | goto send_command_error; | |
416 | } | |
417 | ||
2261ec3d | 418 | iser_task->status = ISER_TASK_STATUS_STARTED; |
e85b24b5 | 419 | |
6df5a128 | 420 | err = iser_post_send(&iser_conn->ib_conn, tx_desc, |
6ec9d4d2 | 421 | iser_signal_comp(sig_count)); |
e85b24b5 OG |
422 | if (!err) |
423 | return 0; | |
424 | ||
425 | send_command_error: | |
2261ec3d | 426 | iser_err("conn %p failed task->itt %d err %d\n",conn, task->itt, err); |
e85b24b5 OG |
427 | return err; |
428 | } | |
429 | ||
430 | /** | |
431 | * iser_send_data_out - send data out PDU | |
432 | */ | |
2747fdb2 | 433 | int iser_send_data_out(struct iscsi_conn *conn, |
2261ec3d | 434 | struct iscsi_task *task, |
e85b24b5 OG |
435 | struct iscsi_data *hdr) |
436 | { | |
5716af6e | 437 | struct iser_conn *iser_conn = conn->dd_data; |
2261ec3d | 438 | struct iscsi_iser_task *iser_task = task->dd_data; |
d769e6be | 439 | struct iser_tx_desc *tx_desc; |
b130eded | 440 | struct iser_mem_reg *mem_reg; |
e85b24b5 OG |
441 | unsigned long buf_offset; |
442 | unsigned long data_seg_len; | |
0a22ab92 | 443 | uint32_t itt; |
d1673905 | 444 | int err; |
f19624aa OG |
445 | struct ib_sge *tx_dsg; |
446 | ||
0a22ab92 | 447 | itt = (__force uint32_t)hdr->itt; |
e85b24b5 OG |
448 | data_seg_len = ntoh24(hdr->dlength); |
449 | buf_offset = ntohl(hdr->offset); | |
450 | ||
451 | iser_dbg("%s itt %d dseg_len %d offset %d\n", | |
452 | __func__,(int)itt,(int)data_seg_len,(int)buf_offset); | |
453 | ||
528f4e8c | 454 | tx_desc = kmem_cache_zalloc(ig.desc_cache, GFP_ATOMIC); |
cac51411 | 455 | if (!tx_desc) |
e85b24b5 | 456 | return -ENOMEM; |
e85b24b5 OG |
457 | |
458 | tx_desc->type = ISCSI_TX_DATAOUT; | |
cfeb91b3 | 459 | tx_desc->cqe.done = iser_dataout_comp; |
f19624aa | 460 | tx_desc->iser_header.flags = ISER_VER; |
e85b24b5 OG |
461 | memcpy(&tx_desc->iscsi_header, hdr, sizeof(struct iscsi_hdr)); |
462 | ||
f19624aa | 463 | /* build the tx desc */ |
d1673905 SG |
464 | err = iser_initialize_task_headers(task, tx_desc); |
465 | if (err) | |
466 | goto send_data_out_error; | |
e85b24b5 | 467 | |
b130eded | 468 | mem_reg = &iser_task->rdma_reg[ISER_DIR_OUT]; |
f19624aa | 469 | tx_dsg = &tx_desc->tx_sg[1]; |
90a6684c SG |
470 | tx_dsg->addr = mem_reg->sge.addr + buf_offset; |
471 | tx_dsg->length = data_seg_len; | |
472 | tx_dsg->lkey = mem_reg->sge.lkey; | |
f19624aa | 473 | tx_desc->num_sge = 2; |
e85b24b5 | 474 | |
2261ec3d | 475 | if (buf_offset + data_seg_len > iser_task->data[ISER_DIR_OUT].data_len) { |
4cb24c55 | 476 | iser_err("Offset:%ld & DSL:%ld in Data-Out inconsistent with total len:%ld, itt:%d\n", |
e85b24b5 | 477 | buf_offset, data_seg_len, |
2261ec3d | 478 | iser_task->data[ISER_DIR_OUT].data_len, itt); |
e85b24b5 OG |
479 | err = -EINVAL; |
480 | goto send_data_out_error; | |
481 | } | |
482 | iser_dbg("data-out itt: %d, offset: %ld, sz: %ld\n", | |
483 | itt, buf_offset, data_seg_len); | |
484 | ||
485 | ||
6df5a128 | 486 | err = iser_post_send(&iser_conn->ib_conn, tx_desc, true); |
e85b24b5 OG |
487 | if (!err) |
488 | return 0; | |
489 | ||
490 | send_data_out_error: | |
e85b24b5 | 491 | kmem_cache_free(ig.desc_cache, tx_desc); |
d1673905 | 492 | iser_err("conn %p failed err %d\n", conn, err); |
e85b24b5 OG |
493 | return err; |
494 | } | |
495 | ||
496 | int iser_send_control(struct iscsi_conn *conn, | |
2261ec3d | 497 | struct iscsi_task *task) |
e85b24b5 | 498 | { |
5716af6e | 499 | struct iser_conn *iser_conn = conn->dd_data; |
2261ec3d | 500 | struct iscsi_iser_task *iser_task = task->dd_data; |
f19624aa | 501 | struct iser_tx_desc *mdesc = &iser_task->desc; |
e85b24b5 | 502 | unsigned long data_seg_len; |
f19624aa | 503 | int err = 0; |
e85b24b5 OG |
504 | struct iser_device *device; |
505 | ||
e85b24b5 OG |
506 | /* build the tx desc regd header and add it to the tx desc dto */ |
507 | mdesc->type = ISCSI_TX_CONTROL; | |
cfeb91b3 | 508 | mdesc->cqe.done = iser_ctrl_comp; |
5716af6e | 509 | iser_create_send_desc(iser_conn, mdesc); |
e85b24b5 | 510 | |
a4ee3539 | 511 | device = iser_conn->ib_conn.device; |
e85b24b5 | 512 | |
2261ec3d | 513 | data_seg_len = ntoh24(task->hdr->dlength); |
e85b24b5 OG |
514 | |
515 | if (data_seg_len > 0) { | |
0f512b34 | 516 | struct iser_login_desc *desc = &iser_conn->login_desc; |
f19624aa | 517 | struct ib_sge *tx_dsg = &mdesc->tx_sg[1]; |
0f512b34 | 518 | |
f19624aa OG |
519 | if (task != conn->login_task) { |
520 | iser_err("data present on non login task!!!\n"); | |
521 | goto send_control_error; | |
522 | } | |
2c4ce609 | 523 | |
0f512b34 SG |
524 | ib_dma_sync_single_for_cpu(device->ib_device, desc->req_dma, |
525 | task->data_count, DMA_TO_DEVICE); | |
2c4ce609 | 526 | |
0f512b34 | 527 | memcpy(desc->req, task->data, task->data_count); |
2c4ce609 | 528 | |
0f512b34 SG |
529 | ib_dma_sync_single_for_device(device->ib_device, desc->req_dma, |
530 | task->data_count, DMA_TO_DEVICE); | |
2c4ce609 | 531 | |
0f512b34 SG |
532 | tx_dsg->addr = desc->req_dma; |
533 | tx_dsg->length = task->data_count; | |
534 | tx_dsg->lkey = device->pd->local_dma_lkey; | |
f19624aa | 535 | mdesc->num_sge = 2; |
e85b24b5 OG |
536 | } |
537 | ||
bcc60c38 | 538 | if (task == conn->login_task) { |
6a06a4b8 OG |
539 | iser_dbg("op %x dsl %lx, posting login rx buffer\n", |
540 | task->hdr->opcode, data_seg_len); | |
5716af6e | 541 | err = iser_post_recvl(iser_conn); |
bcc60c38 OG |
542 | if (err) |
543 | goto send_control_error; | |
89e984e2 OG |
544 | err = iser_post_rx_bufs(conn, task->hdr); |
545 | if (err) | |
546 | goto send_control_error; | |
e85b24b5 OG |
547 | } |
548 | ||
6df5a128 | 549 | err = iser_post_send(&iser_conn->ib_conn, mdesc, true); |
e85b24b5 OG |
550 | if (!err) |
551 | return 0; | |
552 | ||
553 | send_control_error: | |
e85b24b5 OG |
554 | iser_err("conn %p failed err %d\n",conn, err); |
555 | return err; | |
556 | } | |
557 | ||
cfeb91b3 | 558 | void iser_login_rsp(struct ib_cq *cq, struct ib_wc *wc) |
e85b24b5 | 559 | { |
cfeb91b3 | 560 | struct ib_conn *ib_conn = wc->qp->qp_context; |
7edc5a99 | 561 | struct iser_conn *iser_conn = to_iser_conn(ib_conn); |
cfeb91b3 | 562 | struct iser_login_desc *desc = iser_login(wc->wr_cqe); |
e85b24b5 | 563 | struct iscsi_hdr *hdr; |
0f512b34 | 564 | char *data; |
cfeb91b3 CH |
565 | int length; |
566 | ||
567 | if (unlikely(wc->status != IB_WC_SUCCESS)) { | |
568 | iser_err_comp(wc, "login_rsp"); | |
569 | return; | |
570 | } | |
571 | ||
572 | ib_dma_sync_single_for_cpu(ib_conn->device->ib_device, | |
573 | desc->rsp_dma, ISER_RX_LOGIN_SIZE, | |
574 | DMA_FROM_DEVICE); | |
575 | ||
d3cf81f9 | 576 | hdr = desc->rsp + sizeof(struct iser_ctrl); |
cfeb91b3 CH |
577 | data = desc->rsp + ISER_HEADERS_LEN; |
578 | length = wc->byte_len - ISER_HEADERS_LEN; | |
579 | ||
580 | iser_dbg("op 0x%x itt 0x%x dlen %d\n", hdr->opcode, | |
581 | hdr->itt, length); | |
582 | ||
583 | iscsi_iser_recv(iser_conn->iscsi_conn, hdr, data, length); | |
584 | ||
585 | ib_dma_sync_single_for_device(ib_conn->device->ib_device, | |
586 | desc->rsp_dma, ISER_RX_LOGIN_SIZE, | |
587 | DMA_FROM_DEVICE); | |
588 | ||
589 | ib_conn->post_recv_buf_count--; | |
590 | } | |
591 | ||
59caaed7 JD |
592 | static inline void |
593 | iser_inv_desc(struct iser_fr_desc *desc, u32 rkey) | |
594 | { | |
595 | if (likely(rkey == desc->rsc.mr->rkey)) | |
596 | desc->rsc.mr_valid = 0; | |
597 | else if (likely(rkey == desc->pi_ctx->sig_mr->rkey)) | |
598 | desc->pi_ctx->sig_mr_valid = 0; | |
599 | } | |
600 | ||
601 | static int | |
602 | iser_check_remote_inv(struct iser_conn *iser_conn, | |
603 | struct ib_wc *wc, | |
604 | struct iscsi_hdr *hdr) | |
605 | { | |
606 | if (wc->wc_flags & IB_WC_WITH_INVALIDATE) { | |
607 | struct iscsi_task *task; | |
608 | u32 rkey = wc->ex.invalidate_rkey; | |
609 | ||
610 | iser_dbg("conn %p: remote invalidation for rkey %#x\n", | |
611 | iser_conn, rkey); | |
612 | ||
613 | if (unlikely(!iser_conn->snd_w_inv)) { | |
4cb24c55 ME |
614 | iser_err("conn %p: unexpected remote invalidation, terminating connection\n", |
615 | iser_conn); | |
59caaed7 JD |
616 | return -EPROTO; |
617 | } | |
618 | ||
619 | task = iscsi_itt_to_ctask(iser_conn->iscsi_conn, hdr->itt); | |
620 | if (likely(task)) { | |
621 | struct iscsi_iser_task *iser_task = task->dd_data; | |
622 | struct iser_fr_desc *desc; | |
623 | ||
624 | if (iser_task->dir[ISER_DIR_IN]) { | |
625 | desc = iser_task->rdma_reg[ISER_DIR_IN].mem_h; | |
626 | iser_inv_desc(desc, rkey); | |
627 | } | |
628 | ||
629 | if (iser_task->dir[ISER_DIR_OUT]) { | |
630 | desc = iser_task->rdma_reg[ISER_DIR_OUT].mem_h; | |
631 | iser_inv_desc(desc, rkey); | |
632 | } | |
633 | } else { | |
634 | iser_err("failed to get task for itt=%d\n", hdr->itt); | |
635 | return -EINVAL; | |
636 | } | |
637 | } | |
638 | ||
639 | return 0; | |
640 | } | |
641 | ||
642 | ||
cfeb91b3 CH |
643 | void iser_task_rsp(struct ib_cq *cq, struct ib_wc *wc) |
644 | { | |
645 | struct ib_conn *ib_conn = wc->qp->qp_context; | |
646 | struct iser_conn *iser_conn = to_iser_conn(ib_conn); | |
647 | struct iser_rx_desc *desc = iser_rx(wc->wr_cqe); | |
648 | struct iscsi_hdr *hdr; | |
649 | int length; | |
650 | int outstanding, count, err; | |
651 | ||
652 | if (unlikely(wc->status != IB_WC_SUCCESS)) { | |
653 | iser_err_comp(wc, "task_rsp"); | |
654 | return; | |
bcc60c38 | 655 | } |
e85b24b5 | 656 | |
cfeb91b3 CH |
657 | ib_dma_sync_single_for_cpu(ib_conn->device->ib_device, |
658 | desc->dma_addr, ISER_RX_PAYLOAD_SIZE, | |
659 | DMA_FROM_DEVICE); | |
e85b24b5 | 660 | |
cfeb91b3 CH |
661 | hdr = &desc->iscsi_header; |
662 | length = wc->byte_len - ISER_HEADERS_LEN; | |
e85b24b5 | 663 | |
bcc60c38 | 664 | iser_dbg("op 0x%x itt 0x%x dlen %d\n", hdr->opcode, |
cfeb91b3 | 665 | hdr->itt, length); |
e85b24b5 | 666 | |
59caaed7 JD |
667 | if (iser_check_remote_inv(iser_conn, wc, hdr)) { |
668 | iscsi_conn_failure(iser_conn->iscsi_conn, | |
669 | ISCSI_ERR_CONN_FAILED); | |
670 | return; | |
671 | } | |
672 | ||
cfeb91b3 | 673 | iscsi_iser_recv(iser_conn->iscsi_conn, hdr, desc->data, length); |
e85b24b5 | 674 | |
cfeb91b3 CH |
675 | ib_dma_sync_single_for_device(ib_conn->device->ib_device, |
676 | desc->dma_addr, ISER_RX_PAYLOAD_SIZE, | |
677 | DMA_FROM_DEVICE); | |
e85b24b5 OG |
678 | |
679 | /* decrementing conn->post_recv_buf_count only --after-- freeing the * | |
680 | * task eliminates the need to worry on tasks which are completed in * | |
681 | * parallel to the execution of iser_conn_term. So the code that waits * | |
682 | * for the posted rx bufs refcount to become zero handles everything */ | |
a4ee3539 | 683 | ib_conn->post_recv_buf_count--; |
bcc60c38 | 684 | |
a4ee3539 | 685 | outstanding = ib_conn->post_recv_buf_count; |
5716af6e SG |
686 | if (outstanding + iser_conn->min_posted_rx <= iser_conn->qp_max_recv_dtos) { |
687 | count = min(iser_conn->qp_max_recv_dtos - outstanding, | |
688 | iser_conn->min_posted_rx); | |
689 | err = iser_post_recvm(iser_conn, count); | |
bcc60c38 OG |
690 | if (err) |
691 | iser_err("posting %d rx bufs err %d\n", count, err); | |
692 | } | |
e85b24b5 OG |
693 | } |
694 | ||
cfeb91b3 CH |
695 | void iser_cmd_comp(struct ib_cq *cq, struct ib_wc *wc) |
696 | { | |
697 | if (unlikely(wc->status != IB_WC_SUCCESS)) | |
698 | iser_err_comp(wc, "command"); | |
699 | } | |
700 | ||
701 | void iser_ctrl_comp(struct ib_cq *cq, struct ib_wc *wc) | |
e85b24b5 | 702 | { |
cfeb91b3 | 703 | struct iser_tx_desc *desc = iser_tx(wc->wr_cqe); |
2261ec3d | 704 | struct iscsi_task *task; |
e85b24b5 | 705 | |
cfeb91b3 CH |
706 | if (unlikely(wc->status != IB_WC_SUCCESS)) { |
707 | iser_err_comp(wc, "control"); | |
708 | return; | |
f19624aa | 709 | } |
e85b24b5 | 710 | |
cfeb91b3 CH |
711 | /* this arithmetic is legal by libiscsi dd_data allocation */ |
712 | task = (void *)desc - sizeof(struct iscsi_task); | |
713 | if (task->hdr->itt == RESERVED_ITT) | |
714 | iscsi_put_task(task); | |
715 | } | |
716 | ||
717 | void iser_dataout_comp(struct ib_cq *cq, struct ib_wc *wc) | |
718 | { | |
719 | struct iser_tx_desc *desc = iser_tx(wc->wr_cqe); | |
720 | struct ib_conn *ib_conn = wc->qp->qp_context; | |
721 | struct iser_device *device = ib_conn->device; | |
722 | ||
723 | if (unlikely(wc->status != IB_WC_SUCCESS)) | |
724 | iser_err_comp(wc, "dataout"); | |
725 | ||
726 | ib_dma_unmap_single(device->ib_device, desc->dma_addr, | |
727 | ISER_HEADERS_LEN, DMA_TO_DEVICE); | |
728 | kmem_cache_free(ig.desc_cache, desc); | |
729 | } | |
730 | ||
2261ec3d | 731 | void iser_task_rdma_init(struct iscsi_iser_task *iser_task) |
e85b24b5 OG |
732 | |
733 | { | |
2261ec3d | 734 | iser_task->status = ISER_TASK_STATUS_INIT; |
e85b24b5 | 735 | |
2261ec3d MC |
736 | iser_task->dir[ISER_DIR_IN] = 0; |
737 | iser_task->dir[ISER_DIR_OUT] = 0; | |
e85b24b5 | 738 | |
2261ec3d MC |
739 | iser_task->data[ISER_DIR_IN].data_len = 0; |
740 | iser_task->data[ISER_DIR_OUT].data_len = 0; | |
e85b24b5 | 741 | |
177e31bd SG |
742 | iser_task->prot[ISER_DIR_IN].data_len = 0; |
743 | iser_task->prot[ISER_DIR_OUT].data_len = 0; | |
744 | ||
b130eded SG |
745 | memset(&iser_task->rdma_reg[ISER_DIR_IN], 0, |
746 | sizeof(struct iser_mem_reg)); | |
747 | memset(&iser_task->rdma_reg[ISER_DIR_OUT], 0, | |
748 | sizeof(struct iser_mem_reg)); | |
e85b24b5 OG |
749 | } |
750 | ||
2261ec3d | 751 | void iser_task_rdma_finalize(struct iscsi_iser_task *iser_task) |
e85b24b5 | 752 | { |
177e31bd | 753 | int prot_count = scsi_prot_sg_count(iser_task->sc); |
e85b24b5 | 754 | |
9a8b08fa | 755 | if (iser_task->dir[ISER_DIR_IN]) { |
32467c42 | 756 | iser_unreg_rdma_mem(iser_task, ISER_DIR_IN); |
dd0107a0 SG |
757 | iser_dma_unmap_task_data(iser_task, |
758 | &iser_task->data[ISER_DIR_IN], | |
759 | DMA_FROM_DEVICE); | |
760 | if (prot_count) | |
177e31bd | 761 | iser_dma_unmap_task_data(iser_task, |
c6c95ef4 RD |
762 | &iser_task->prot[ISER_DIR_IN], |
763 | DMA_FROM_DEVICE); | |
9a8b08fa | 764 | } |
e85b24b5 | 765 | |
9a8b08fa | 766 | if (iser_task->dir[ISER_DIR_OUT]) { |
32467c42 | 767 | iser_unreg_rdma_mem(iser_task, ISER_DIR_OUT); |
dd0107a0 SG |
768 | iser_dma_unmap_task_data(iser_task, |
769 | &iser_task->data[ISER_DIR_OUT], | |
770 | DMA_TO_DEVICE); | |
771 | if (prot_count) | |
9a8b08fa | 772 | iser_dma_unmap_task_data(iser_task, |
c6c95ef4 RD |
773 | &iser_task->prot[ISER_DIR_OUT], |
774 | DMA_TO_DEVICE); | |
9a8b08fa | 775 | } |
e85b24b5 | 776 | } |