| 1 | // SPDX-License-Identifier: GPL-2.0 |
| 2 | /* |
| 3 | * qdio queue initialization |
| 4 | * |
| 5 | * Copyright IBM Corp. 2008 |
| 6 | * Author(s): Jan Glauber <jang@linux.vnet.ibm.com> |
| 7 | */ |
| 8 | #include <linux/kernel.h> |
| 9 | #include <linux/slab.h> |
| 10 | #include <linux/export.h> |
| 11 | #include <linux/io.h> |
| 12 | #include <asm/qdio.h> |
| 13 | |
| 14 | #include "cio.h" |
| 15 | #include "css.h" |
| 16 | #include "device.h" |
| 17 | #include "ioasm.h" |
| 18 | #include "chsc.h" |
| 19 | #include "qdio.h" |
| 20 | #include "qdio_debug.h" |
| 21 | |
| 22 | #define QBUFF_PER_PAGE (PAGE_SIZE / sizeof(struct qdio_buffer)) |
| 23 | |
| 24 | static struct kmem_cache *qdio_q_cache; |
| 25 | static struct kmem_cache *qdio_aob_cache; |
| 26 | |
| 27 | struct qaob *qdio_allocate_aob(void) |
| 28 | { |
| 29 | return kmem_cache_zalloc(qdio_aob_cache, GFP_ATOMIC); |
| 30 | } |
| 31 | |
| 32 | void qdio_release_aob(struct qaob *aob) |
| 33 | { |
| 34 | kmem_cache_free(qdio_aob_cache, aob); |
| 35 | } |
| 36 | EXPORT_SYMBOL_GPL(qdio_release_aob); |
| 37 | |
| 38 | /** |
| 39 | * qdio_free_buffers() - free qdio buffers |
| 40 | * @buf: array of pointers to qdio buffers |
| 41 | * @count: number of qdio buffers to free |
| 42 | */ |
| 43 | void qdio_free_buffers(struct qdio_buffer **buf, unsigned int count) |
| 44 | { |
| 45 | int pos; |
| 46 | |
| 47 | for (pos = 0; pos < count; pos += QBUFF_PER_PAGE) |
| 48 | free_page((unsigned long) buf[pos]); |
| 49 | } |
| 50 | EXPORT_SYMBOL_GPL(qdio_free_buffers); |
| 51 | |
| 52 | /** |
| 53 | * qdio_alloc_buffers() - allocate qdio buffers |
| 54 | * @buf: array of pointers to qdio buffers |
| 55 | * @count: number of qdio buffers to allocate |
| 56 | */ |
| 57 | int qdio_alloc_buffers(struct qdio_buffer **buf, unsigned int count) |
| 58 | { |
| 59 | int pos; |
| 60 | |
| 61 | for (pos = 0; pos < count; pos += QBUFF_PER_PAGE) { |
| 62 | buf[pos] = (void *) get_zeroed_page(GFP_KERNEL); |
| 63 | if (!buf[pos]) { |
| 64 | qdio_free_buffers(buf, count); |
| 65 | return -ENOMEM; |
| 66 | } |
| 67 | } |
| 68 | for (pos = 0; pos < count; pos++) |
| 69 | if (pos % QBUFF_PER_PAGE) |
| 70 | buf[pos] = buf[pos - 1] + 1; |
| 71 | return 0; |
| 72 | } |
| 73 | EXPORT_SYMBOL_GPL(qdio_alloc_buffers); |
| 74 | |
| 75 | /** |
| 76 | * qdio_reset_buffers() - reset qdio buffers |
| 77 | * @buf: array of pointers to qdio buffers |
| 78 | * @count: number of qdio buffers that will be zeroed |
| 79 | */ |
| 80 | void qdio_reset_buffers(struct qdio_buffer **buf, unsigned int count) |
| 81 | { |
| 82 | int pos; |
| 83 | |
| 84 | for (pos = 0; pos < count; pos++) |
| 85 | memset(buf[pos], 0, sizeof(struct qdio_buffer)); |
| 86 | } |
| 87 | EXPORT_SYMBOL_GPL(qdio_reset_buffers); |
| 88 | |
| 89 | /* |
| 90 | * qebsm is only available under 64bit but the adapter sets the feature |
| 91 | * flag anyway, so we manually override it. |
| 92 | */ |
| 93 | static inline int qebsm_possible(void) |
| 94 | { |
| 95 | return css_general_characteristics.qebsm; |
| 96 | } |
| 97 | |
| 98 | /* |
| 99 | * qib_param_field: pointer to 128 bytes or NULL, if no param field |
| 100 | * nr_input_qs: pointer to nr_queues*128 words of data or NULL |
| 101 | */ |
| 102 | static void set_impl_params(struct qdio_irq *irq_ptr, |
| 103 | unsigned int qib_param_field_format, |
| 104 | unsigned char *qib_param_field, |
| 105 | unsigned long *input_slib_elements, |
| 106 | unsigned long *output_slib_elements) |
| 107 | { |
| 108 | struct qdio_q *q; |
| 109 | int i, j; |
| 110 | |
| 111 | if (!irq_ptr) |
| 112 | return; |
| 113 | |
| 114 | irq_ptr->qib.pfmt = qib_param_field_format; |
| 115 | if (qib_param_field) |
| 116 | memcpy(irq_ptr->qib.parm, qib_param_field, |
| 117 | sizeof(irq_ptr->qib.parm)); |
| 118 | |
| 119 | if (!input_slib_elements) |
| 120 | goto output; |
| 121 | |
| 122 | for_each_input_queue(irq_ptr, q, i) { |
| 123 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) |
| 124 | q->slib->slibe[j].parms = |
| 125 | input_slib_elements[i * QDIO_MAX_BUFFERS_PER_Q + j]; |
| 126 | } |
| 127 | output: |
| 128 | if (!output_slib_elements) |
| 129 | return; |
| 130 | |
| 131 | for_each_output_queue(irq_ptr, q, i) { |
| 132 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) |
| 133 | q->slib->slibe[j].parms = |
| 134 | output_slib_elements[i * QDIO_MAX_BUFFERS_PER_Q + j]; |
| 135 | } |
| 136 | } |
| 137 | |
| 138 | static int __qdio_allocate_qs(struct qdio_q **irq_ptr_qs, int nr_queues) |
| 139 | { |
| 140 | struct qdio_q *q; |
| 141 | int i; |
| 142 | |
| 143 | for (i = 0; i < nr_queues; i++) { |
| 144 | q = kmem_cache_zalloc(qdio_q_cache, GFP_KERNEL); |
| 145 | if (!q) |
| 146 | return -ENOMEM; |
| 147 | |
| 148 | q->slib = (struct slib *) __get_free_page(GFP_KERNEL); |
| 149 | if (!q->slib) { |
| 150 | kmem_cache_free(qdio_q_cache, q); |
| 151 | return -ENOMEM; |
| 152 | } |
| 153 | irq_ptr_qs[i] = q; |
| 154 | } |
| 155 | return 0; |
| 156 | } |
| 157 | |
| 158 | int qdio_allocate_qs(struct qdio_irq *irq_ptr, int nr_input_qs, int nr_output_qs) |
| 159 | { |
| 160 | int rc; |
| 161 | |
| 162 | rc = __qdio_allocate_qs(irq_ptr->input_qs, nr_input_qs); |
| 163 | if (rc) |
| 164 | return rc; |
| 165 | rc = __qdio_allocate_qs(irq_ptr->output_qs, nr_output_qs); |
| 166 | return rc; |
| 167 | } |
| 168 | |
| 169 | static void setup_queues_misc(struct qdio_q *q, struct qdio_irq *irq_ptr, |
| 170 | qdio_handler_t *handler, int i) |
| 171 | { |
| 172 | struct slib *slib = q->slib; |
| 173 | |
| 174 | /* queue must be cleared for qdio_establish */ |
| 175 | memset(q, 0, sizeof(*q)); |
| 176 | memset(slib, 0, PAGE_SIZE); |
| 177 | q->slib = slib; |
| 178 | q->irq_ptr = irq_ptr; |
| 179 | q->mask = 1 << (31 - i); |
| 180 | q->nr = i; |
| 181 | q->handler = handler; |
| 182 | } |
| 183 | |
| 184 | static void setup_storage_lists(struct qdio_q *q, struct qdio_irq *irq_ptr, |
| 185 | struct qdio_buffer **sbals_array, int i) |
| 186 | { |
| 187 | struct qdio_q *prev; |
| 188 | int j; |
| 189 | |
| 190 | DBF_HEX(&q, sizeof(void *)); |
| 191 | q->sl = (struct sl *)((char *)q->slib + PAGE_SIZE / 2); |
| 192 | |
| 193 | /* fill in sbal */ |
| 194 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) |
| 195 | q->sbal[j] = *sbals_array++; |
| 196 | |
| 197 | /* fill in slib */ |
| 198 | if (i > 0) { |
| 199 | prev = (q->is_input_q) ? irq_ptr->input_qs[i - 1] |
| 200 | : irq_ptr->output_qs[i - 1]; |
| 201 | prev->slib->nsliba = (unsigned long)q->slib; |
| 202 | } |
| 203 | |
| 204 | q->slib->sla = (unsigned long)q->sl; |
| 205 | q->slib->slsba = (unsigned long)&q->slsb.val[0]; |
| 206 | |
| 207 | /* fill in sl */ |
| 208 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) |
| 209 | q->sl->element[j].sbal = virt_to_phys(q->sbal[j]); |
| 210 | } |
| 211 | |
| 212 | static void setup_queues(struct qdio_irq *irq_ptr, |
| 213 | struct qdio_initialize *qdio_init) |
| 214 | { |
| 215 | struct qdio_q *q; |
| 216 | struct qdio_outbuf_state *output_sbal_state_array = |
| 217 | qdio_init->output_sbal_state_array; |
| 218 | int i; |
| 219 | |
| 220 | for_each_input_queue(irq_ptr, q, i) { |
| 221 | DBF_EVENT("inq:%1d", i); |
| 222 | setup_queues_misc(q, irq_ptr, qdio_init->input_handler, i); |
| 223 | |
| 224 | q->is_input_q = 1; |
| 225 | |
| 226 | setup_storage_lists(q, irq_ptr, |
| 227 | qdio_init->input_sbal_addr_array[i], i); |
| 228 | |
| 229 | if (is_thinint_irq(irq_ptr)) { |
| 230 | tasklet_init(&q->tasklet, tiqdio_inbound_processing, |
| 231 | (unsigned long) q); |
| 232 | } else { |
| 233 | tasklet_init(&q->tasklet, qdio_inbound_processing, |
| 234 | (unsigned long) q); |
| 235 | } |
| 236 | } |
| 237 | |
| 238 | for_each_output_queue(irq_ptr, q, i) { |
| 239 | DBF_EVENT("outq:%1d", i); |
| 240 | setup_queues_misc(q, irq_ptr, qdio_init->output_handler, i); |
| 241 | |
| 242 | q->u.out.sbal_state = output_sbal_state_array; |
| 243 | output_sbal_state_array += QDIO_MAX_BUFFERS_PER_Q; |
| 244 | |
| 245 | q->is_input_q = 0; |
| 246 | setup_storage_lists(q, irq_ptr, |
| 247 | qdio_init->output_sbal_addr_array[i], i); |
| 248 | |
| 249 | tasklet_init(&q->tasklet, qdio_outbound_processing, |
| 250 | (unsigned long) q); |
| 251 | timer_setup(&q->u.out.timer, qdio_outbound_timer, 0); |
| 252 | } |
| 253 | } |
| 254 | |
| 255 | static void process_ac_flags(struct qdio_irq *irq_ptr, unsigned char qdioac) |
| 256 | { |
| 257 | if (qdioac & AC1_SIGA_INPUT_NEEDED) |
| 258 | irq_ptr->siga_flag.input = 1; |
| 259 | if (qdioac & AC1_SIGA_OUTPUT_NEEDED) |
| 260 | irq_ptr->siga_flag.output = 1; |
| 261 | if (qdioac & AC1_SIGA_SYNC_NEEDED) |
| 262 | irq_ptr->siga_flag.sync = 1; |
| 263 | if (!(qdioac & AC1_AUTOMATIC_SYNC_ON_THININT)) |
| 264 | irq_ptr->siga_flag.sync_after_ai = 1; |
| 265 | if (!(qdioac & AC1_AUTOMATIC_SYNC_ON_OUT_PCI)) |
| 266 | irq_ptr->siga_flag.sync_out_after_pci = 1; |
| 267 | } |
| 268 | |
| 269 | static void check_and_setup_qebsm(struct qdio_irq *irq_ptr, |
| 270 | unsigned char qdioac, unsigned long token) |
| 271 | { |
| 272 | if (!(irq_ptr->qib.rflags & QIB_RFLAGS_ENABLE_QEBSM)) |
| 273 | goto no_qebsm; |
| 274 | if (!(qdioac & AC1_SC_QEBSM_AVAILABLE) || |
| 275 | (!(qdioac & AC1_SC_QEBSM_ENABLED))) |
| 276 | goto no_qebsm; |
| 277 | |
| 278 | irq_ptr->sch_token = token; |
| 279 | |
| 280 | DBF_EVENT("V=V:1"); |
| 281 | DBF_EVENT("%8lx", irq_ptr->sch_token); |
| 282 | return; |
| 283 | |
| 284 | no_qebsm: |
| 285 | irq_ptr->sch_token = 0; |
| 286 | irq_ptr->qib.rflags &= ~QIB_RFLAGS_ENABLE_QEBSM; |
| 287 | DBF_EVENT("noV=V"); |
| 288 | } |
| 289 | |
| 290 | /* |
| 291 | * If there is a qdio_irq we use the chsc_page and store the information |
| 292 | * in the qdio_irq, otherwise we copy it to the specified structure. |
| 293 | */ |
| 294 | int qdio_setup_get_ssqd(struct qdio_irq *irq_ptr, |
| 295 | struct subchannel_id *schid, |
| 296 | struct qdio_ssqd_desc *data) |
| 297 | { |
| 298 | struct chsc_ssqd_area *ssqd; |
| 299 | int rc; |
| 300 | |
| 301 | DBF_EVENT("getssqd:%4x", schid->sch_no); |
| 302 | if (!irq_ptr) { |
| 303 | ssqd = (struct chsc_ssqd_area *)__get_free_page(GFP_KERNEL); |
| 304 | if (!ssqd) |
| 305 | return -ENOMEM; |
| 306 | } else { |
| 307 | ssqd = (struct chsc_ssqd_area *)irq_ptr->chsc_page; |
| 308 | } |
| 309 | |
| 310 | rc = chsc_ssqd(*schid, ssqd); |
| 311 | if (rc) |
| 312 | goto out; |
| 313 | |
| 314 | if (!(ssqd->qdio_ssqd.flags & CHSC_FLAG_QDIO_CAPABILITY) || |
| 315 | !(ssqd->qdio_ssqd.flags & CHSC_FLAG_VALIDITY) || |
| 316 | (ssqd->qdio_ssqd.sch != schid->sch_no)) |
| 317 | rc = -EINVAL; |
| 318 | |
| 319 | if (!rc) |
| 320 | memcpy(data, &ssqd->qdio_ssqd, sizeof(*data)); |
| 321 | |
| 322 | out: |
| 323 | if (!irq_ptr) |
| 324 | free_page((unsigned long)ssqd); |
| 325 | |
| 326 | return rc; |
| 327 | } |
| 328 | |
| 329 | void qdio_setup_ssqd_info(struct qdio_irq *irq_ptr) |
| 330 | { |
| 331 | unsigned char qdioac; |
| 332 | int rc; |
| 333 | |
| 334 | rc = qdio_setup_get_ssqd(irq_ptr, &irq_ptr->schid, &irq_ptr->ssqd_desc); |
| 335 | if (rc) { |
| 336 | DBF_ERROR("%4x ssqd ERR", irq_ptr->schid.sch_no); |
| 337 | DBF_ERROR("rc:%x", rc); |
| 338 | /* all flags set, worst case */ |
| 339 | qdioac = AC1_SIGA_INPUT_NEEDED | AC1_SIGA_OUTPUT_NEEDED | |
| 340 | AC1_SIGA_SYNC_NEEDED; |
| 341 | } else |
| 342 | qdioac = irq_ptr->ssqd_desc.qdioac1; |
| 343 | |
| 344 | check_and_setup_qebsm(irq_ptr, qdioac, irq_ptr->ssqd_desc.sch_token); |
| 345 | process_ac_flags(irq_ptr, qdioac); |
| 346 | DBF_EVENT("ac 1:%2x 2:%4x", qdioac, irq_ptr->ssqd_desc.qdioac2); |
| 347 | DBF_EVENT("3:%4x qib:%4x", irq_ptr->ssqd_desc.qdioac3, irq_ptr->qib.ac); |
| 348 | } |
| 349 | |
| 350 | void qdio_release_memory(struct qdio_irq *irq_ptr) |
| 351 | { |
| 352 | struct qdio_q *q; |
| 353 | int i; |
| 354 | |
| 355 | /* |
| 356 | * Must check queue array manually since irq_ptr->nr_input_queues / |
| 357 | * irq_ptr->nr_input_queues may not yet be set. |
| 358 | */ |
| 359 | for (i = 0; i < QDIO_MAX_QUEUES_PER_IRQ; i++) { |
| 360 | q = irq_ptr->input_qs[i]; |
| 361 | if (q) { |
| 362 | free_page((unsigned long) q->slib); |
| 363 | kmem_cache_free(qdio_q_cache, q); |
| 364 | } |
| 365 | } |
| 366 | for (i = 0; i < QDIO_MAX_QUEUES_PER_IRQ; i++) { |
| 367 | q = irq_ptr->output_qs[i]; |
| 368 | if (q) { |
| 369 | if (q->u.out.use_cq) { |
| 370 | int n; |
| 371 | |
| 372 | for (n = 0; n < QDIO_MAX_BUFFERS_PER_Q; ++n) { |
| 373 | struct qaob *aob = q->u.out.aobs[n]; |
| 374 | if (aob) { |
| 375 | qdio_release_aob(aob); |
| 376 | q->u.out.aobs[n] = NULL; |
| 377 | } |
| 378 | } |
| 379 | |
| 380 | qdio_disable_async_operation(&q->u.out); |
| 381 | } |
| 382 | free_page((unsigned long) q->slib); |
| 383 | kmem_cache_free(qdio_q_cache, q); |
| 384 | } |
| 385 | } |
| 386 | free_page((unsigned long) irq_ptr->qdr); |
| 387 | free_page(irq_ptr->chsc_page); |
| 388 | free_page((unsigned long) irq_ptr); |
| 389 | } |
| 390 | |
| 391 | static void __qdio_allocate_fill_qdr(struct qdio_irq *irq_ptr, |
| 392 | struct qdio_q **irq_ptr_qs, |
| 393 | int i, int nr) |
| 394 | { |
| 395 | irq_ptr->qdr->qdf0[i + nr].sliba = |
| 396 | (unsigned long)irq_ptr_qs[i]->slib; |
| 397 | |
| 398 | irq_ptr->qdr->qdf0[i + nr].sla = |
| 399 | (unsigned long)irq_ptr_qs[i]->sl; |
| 400 | |
| 401 | irq_ptr->qdr->qdf0[i + nr].slsba = |
| 402 | (unsigned long)&irq_ptr_qs[i]->slsb.val[0]; |
| 403 | |
| 404 | irq_ptr->qdr->qdf0[i + nr].akey = PAGE_DEFAULT_KEY >> 4; |
| 405 | irq_ptr->qdr->qdf0[i + nr].bkey = PAGE_DEFAULT_KEY >> 4; |
| 406 | irq_ptr->qdr->qdf0[i + nr].ckey = PAGE_DEFAULT_KEY >> 4; |
| 407 | irq_ptr->qdr->qdf0[i + nr].dkey = PAGE_DEFAULT_KEY >> 4; |
| 408 | } |
| 409 | |
| 410 | static void setup_qdr(struct qdio_irq *irq_ptr, |
| 411 | struct qdio_initialize *qdio_init) |
| 412 | { |
| 413 | int i; |
| 414 | |
| 415 | irq_ptr->qdr->qfmt = qdio_init->q_format; |
| 416 | irq_ptr->qdr->ac = qdio_init->qdr_ac; |
| 417 | irq_ptr->qdr->iqdcnt = qdio_init->no_input_qs; |
| 418 | irq_ptr->qdr->oqdcnt = qdio_init->no_output_qs; |
| 419 | irq_ptr->qdr->iqdsz = sizeof(struct qdesfmt0) / 4; /* size in words */ |
| 420 | irq_ptr->qdr->oqdsz = sizeof(struct qdesfmt0) / 4; |
| 421 | irq_ptr->qdr->qiba = (unsigned long)&irq_ptr->qib; |
| 422 | irq_ptr->qdr->qkey = PAGE_DEFAULT_KEY >> 4; |
| 423 | |
| 424 | for (i = 0; i < qdio_init->no_input_qs; i++) |
| 425 | __qdio_allocate_fill_qdr(irq_ptr, irq_ptr->input_qs, i, 0); |
| 426 | |
| 427 | for (i = 0; i < qdio_init->no_output_qs; i++) |
| 428 | __qdio_allocate_fill_qdr(irq_ptr, irq_ptr->output_qs, i, |
| 429 | qdio_init->no_input_qs); |
| 430 | } |
| 431 | |
| 432 | static void setup_qib(struct qdio_irq *irq_ptr, |
| 433 | struct qdio_initialize *init_data) |
| 434 | { |
| 435 | if (qebsm_possible()) |
| 436 | irq_ptr->qib.rflags |= QIB_RFLAGS_ENABLE_QEBSM; |
| 437 | |
| 438 | irq_ptr->qib.rflags |= init_data->qib_rflags; |
| 439 | |
| 440 | irq_ptr->qib.qfmt = init_data->q_format; |
| 441 | if (init_data->no_input_qs) |
| 442 | irq_ptr->qib.isliba = |
| 443 | (unsigned long)(irq_ptr->input_qs[0]->slib); |
| 444 | if (init_data->no_output_qs) |
| 445 | irq_ptr->qib.osliba = |
| 446 | (unsigned long)(irq_ptr->output_qs[0]->slib); |
| 447 | memcpy(irq_ptr->qib.ebcnam, init_data->adapter_name, 8); |
| 448 | } |
| 449 | |
| 450 | int qdio_setup_irq(struct qdio_irq *irq_ptr, struct qdio_initialize *init_data) |
| 451 | { |
| 452 | struct ccw_device *cdev = irq_ptr->cdev; |
| 453 | struct ciw *ciw; |
| 454 | |
| 455 | memset(&irq_ptr->qib, 0, sizeof(irq_ptr->qib)); |
| 456 | memset(&irq_ptr->siga_flag, 0, sizeof(irq_ptr->siga_flag)); |
| 457 | memset(&irq_ptr->ccw, 0, sizeof(irq_ptr->ccw)); |
| 458 | memset(&irq_ptr->ssqd_desc, 0, sizeof(irq_ptr->ssqd_desc)); |
| 459 | memset(&irq_ptr->perf_stat, 0, sizeof(irq_ptr->perf_stat)); |
| 460 | |
| 461 | irq_ptr->debugfs_dev = NULL; |
| 462 | irq_ptr->sch_token = irq_ptr->perf_stat_enabled = 0; |
| 463 | irq_ptr->state = QDIO_IRQ_STATE_INACTIVE; |
| 464 | |
| 465 | /* wipes qib.ac, required by ar7063 */ |
| 466 | memset(irq_ptr->qdr, 0, sizeof(struct qdr)); |
| 467 | |
| 468 | irq_ptr->int_parm = init_data->int_parm; |
| 469 | irq_ptr->nr_input_qs = init_data->no_input_qs; |
| 470 | irq_ptr->nr_output_qs = init_data->no_output_qs; |
| 471 | irq_ptr->scan_threshold = init_data->scan_threshold; |
| 472 | ccw_device_get_schid(cdev, &irq_ptr->schid); |
| 473 | setup_queues(irq_ptr, init_data); |
| 474 | |
| 475 | if (init_data->irq_poll) { |
| 476 | irq_ptr->irq_poll = init_data->irq_poll; |
| 477 | set_bit(QDIO_IRQ_DISABLED, &irq_ptr->poll_state); |
| 478 | } else { |
| 479 | irq_ptr->irq_poll = NULL; |
| 480 | } |
| 481 | |
| 482 | setup_qib(irq_ptr, init_data); |
| 483 | qdio_setup_thinint(irq_ptr); |
| 484 | set_impl_params(irq_ptr, init_data->qib_param_field_format, |
| 485 | init_data->qib_param_field, |
| 486 | init_data->input_slib_elements, |
| 487 | init_data->output_slib_elements); |
| 488 | |
| 489 | /* fill input and output descriptors */ |
| 490 | setup_qdr(irq_ptr, init_data); |
| 491 | |
| 492 | /* qdr, qib, sls, slsbs, slibs, sbales are filled now */ |
| 493 | |
| 494 | /* get qdio commands */ |
| 495 | ciw = ccw_device_get_ciw(cdev, CIW_TYPE_EQUEUE); |
| 496 | if (!ciw) { |
| 497 | DBF_ERROR("%4x NO EQ", irq_ptr->schid.sch_no); |
| 498 | return -EINVAL; |
| 499 | } |
| 500 | irq_ptr->equeue = *ciw; |
| 501 | |
| 502 | ciw = ccw_device_get_ciw(cdev, CIW_TYPE_AQUEUE); |
| 503 | if (!ciw) { |
| 504 | DBF_ERROR("%4x NO AQ", irq_ptr->schid.sch_no); |
| 505 | return -EINVAL; |
| 506 | } |
| 507 | irq_ptr->aqueue = *ciw; |
| 508 | |
| 509 | /* set new interrupt handler */ |
| 510 | spin_lock_irq(get_ccwdev_lock(cdev)); |
| 511 | irq_ptr->orig_handler = cdev->handler; |
| 512 | cdev->handler = qdio_int_handler; |
| 513 | spin_unlock_irq(get_ccwdev_lock(cdev)); |
| 514 | return 0; |
| 515 | } |
| 516 | |
| 517 | void qdio_print_subchannel_info(struct qdio_irq *irq_ptr) |
| 518 | { |
| 519 | char s[80]; |
| 520 | |
| 521 | snprintf(s, 80, "qdio: %s %s on SC %x using " |
| 522 | "AI:%d QEBSM:%d PRI:%d TDD:%d SIGA:%s%s%s%s%s\n", |
| 523 | dev_name(&irq_ptr->cdev->dev), |
| 524 | (irq_ptr->qib.qfmt == QDIO_QETH_QFMT) ? "OSA" : |
| 525 | ((irq_ptr->qib.qfmt == QDIO_ZFCP_QFMT) ? "ZFCP" : "HS"), |
| 526 | irq_ptr->schid.sch_no, |
| 527 | is_thinint_irq(irq_ptr), |
| 528 | (irq_ptr->sch_token) ? 1 : 0, |
| 529 | pci_out_supported(irq_ptr) ? 1 : 0, |
| 530 | css_general_characteristics.aif_tdd, |
| 531 | (irq_ptr->siga_flag.input) ? "R" : " ", |
| 532 | (irq_ptr->siga_flag.output) ? "W" : " ", |
| 533 | (irq_ptr->siga_flag.sync) ? "S" : " ", |
| 534 | (irq_ptr->siga_flag.sync_after_ai) ? "A" : " ", |
| 535 | (irq_ptr->siga_flag.sync_out_after_pci) ? "P" : " "); |
| 536 | printk(KERN_INFO "%s", s); |
| 537 | } |
| 538 | |
| 539 | int qdio_enable_async_operation(struct qdio_output_q *outq) |
| 540 | { |
| 541 | outq->aobs = kcalloc(QDIO_MAX_BUFFERS_PER_Q, sizeof(struct qaob *), |
| 542 | GFP_KERNEL); |
| 543 | if (!outq->aobs) { |
| 544 | outq->use_cq = 0; |
| 545 | return -ENOMEM; |
| 546 | } |
| 547 | outq->use_cq = 1; |
| 548 | return 0; |
| 549 | } |
| 550 | |
| 551 | void qdio_disable_async_operation(struct qdio_output_q *q) |
| 552 | { |
| 553 | kfree(q->aobs); |
| 554 | q->aobs = NULL; |
| 555 | q->use_cq = 0; |
| 556 | } |
| 557 | |
| 558 | int __init qdio_setup_init(void) |
| 559 | { |
| 560 | int rc; |
| 561 | |
| 562 | qdio_q_cache = kmem_cache_create("qdio_q", sizeof(struct qdio_q), |
| 563 | 256, 0, NULL); |
| 564 | if (!qdio_q_cache) |
| 565 | return -ENOMEM; |
| 566 | |
| 567 | qdio_aob_cache = kmem_cache_create("qdio_aob", |
| 568 | sizeof(struct qaob), |
| 569 | sizeof(struct qaob), |
| 570 | 0, |
| 571 | NULL); |
| 572 | if (!qdio_aob_cache) { |
| 573 | rc = -ENOMEM; |
| 574 | goto free_qdio_q_cache; |
| 575 | } |
| 576 | |
| 577 | /* Check for OSA/FCP thin interrupts (bit 67). */ |
| 578 | DBF_EVENT("thinint:%1d", |
| 579 | (css_general_characteristics.aif_osa) ? 1 : 0); |
| 580 | |
| 581 | /* Check for QEBSM support in general (bit 58). */ |
| 582 | DBF_EVENT("cssQEBSM:%1d", (qebsm_possible()) ? 1 : 0); |
| 583 | rc = 0; |
| 584 | out: |
| 585 | return rc; |
| 586 | free_qdio_q_cache: |
| 587 | kmem_cache_destroy(qdio_q_cache); |
| 588 | goto out; |
| 589 | } |
| 590 | |
| 591 | void qdio_setup_exit(void) |
| 592 | { |
| 593 | kmem_cache_destroy(qdio_aob_cache); |
| 594 | kmem_cache_destroy(qdio_q_cache); |
| 595 | } |