// SPDX-License-Identifier: GPL-2.0 /* Copyright (C) 2018-2025, Advanced Micro Devices, Inc. */ #include #include #include #include "ionic_fw.h" #include "ionic_ibdev.h" #define IONIC_EQ_COUNT_MIN 4 #define IONIC_AQ_COUNT_MIN 1 /* not a valid queue position or negative error status */ #define IONIC_ADMIN_POSTED 0x10000 /* cpu can be held with irq disabled for COUNT * MS (for create/destroy_ah) */ #define IONIC_ADMIN_BUSY_RETRY_COUNT 2000 #define IONIC_ADMIN_BUSY_RETRY_MS 1 /* admin queue will be considered failed if a command takes longer */ #define IONIC_ADMIN_TIMEOUT (HZ * 2) #define IONIC_ADMIN_WARN (HZ / 8) /* will poll for admin cq to tolerate and report from missed event */ #define IONIC_ADMIN_DELAY (HZ / 8) /* work queue for polling the event queue and admin cq */ struct workqueue_struct *ionic_evt_workq; static void ionic_admin_timedout(struct ionic_aq *aq) { struct ionic_ibdev *dev = aq->dev; unsigned long irqflags; u16 pos; spin_lock_irqsave(&aq->lock, irqflags); if (ionic_queue_empty(&aq->q)) goto out; /* Reset ALL adminq if any one times out */ if (atomic_read(&aq->admin_state) < IONIC_ADMIN_KILLED) queue_work(ionic_evt_workq, &dev->reset_work); ibdev_err(&dev->ibdev, "admin command timed out, aq %d after: %ums\n", aq->aqid, (u32)jiffies_to_msecs(jiffies - aq->stamp)); pos = (aq->q.prod - 1) & aq->q.mask; if (pos == aq->q.cons) goto out; ibdev_warn(&dev->ibdev, "admin pos %u (last posted)\n", pos); print_hex_dump(KERN_WARNING, "cmd ", DUMP_PREFIX_OFFSET, 16, 1, ionic_queue_at(&aq->q, pos), BIT(aq->q.stride_log2), true); out: spin_unlock_irqrestore(&aq->lock, irqflags); } static void ionic_admin_reset_dwork(struct ionic_ibdev *dev) { if (atomic_read(&dev->admin_state) == IONIC_ADMIN_KILLED) return; queue_delayed_work(ionic_evt_workq, &dev->admin_dwork, IONIC_ADMIN_DELAY); } static void ionic_admin_reset_wdog(struct ionic_aq *aq) { if (atomic_read(&aq->admin_state) == IONIC_ADMIN_KILLED) return; aq->stamp = jiffies; ionic_admin_reset_dwork(aq->dev); } static bool ionic_admin_next_cqe(struct ionic_ibdev *dev, struct ionic_cq *cq, struct ionic_v1_cqe **cqe) { struct ionic_v1_cqe *qcqe = ionic_queue_at_prod(&cq->q); if (unlikely(cq->color != ionic_v1_cqe_color(qcqe))) return false; /* Prevent out-of-order reads of the CQE */ dma_rmb(); *cqe = qcqe; return true; } static void ionic_admin_poll_locked(struct ionic_aq *aq) { struct ionic_cq *cq = &aq->vcq->cq[0]; struct ionic_admin_wr *wr, *wr_next; struct ionic_ibdev *dev = aq->dev; u32 wr_strides, avlbl_strides; struct ionic_v1_cqe *cqe; u32 qtf, qid; u16 old_prod; u8 type; lockdep_assert_held(&aq->lock); if (atomic_read(&aq->admin_state) == IONIC_ADMIN_KILLED) { list_for_each_entry_safe(wr, wr_next, &aq->wr_prod, aq_ent) { INIT_LIST_HEAD(&wr->aq_ent); aq->q_wr[wr->status].wr = NULL; wr->status = atomic_read(&aq->admin_state); complete_all(&wr->work); } INIT_LIST_HEAD(&aq->wr_prod); list_for_each_entry_safe(wr, wr_next, &aq->wr_post, aq_ent) { INIT_LIST_HEAD(&wr->aq_ent); wr->status = atomic_read(&aq->admin_state); complete_all(&wr->work); } INIT_LIST_HEAD(&aq->wr_post); return; } old_prod = cq->q.prod; while (ionic_admin_next_cqe(dev, cq, &cqe)) { qtf = ionic_v1_cqe_qtf(cqe); qid = ionic_v1_cqe_qtf_qid(qtf); type = ionic_v1_cqe_qtf_type(qtf); if (unlikely(type != IONIC_V1_CQE_TYPE_ADMIN)) { ibdev_warn_ratelimited(&dev->ibdev, "bad cqe type %u\n", type); goto cq_next; } if (unlikely(qid != aq->aqid)) { ibdev_warn_ratelimited(&dev->ibdev, "bad cqe qid %u\n", qid); goto cq_next; } if (unlikely(be16_to_cpu(cqe->admin.cmd_idx) != aq->q.cons)) { ibdev_warn_ratelimited(&dev->ibdev, "bad idx %u cons %u qid %u\n", be16_to_cpu(cqe->admin.cmd_idx), aq->q.cons, qid); goto cq_next; } if (unlikely(ionic_queue_empty(&aq->q))) { ibdev_warn_ratelimited(&dev->ibdev, "bad cqe for empty adminq\n"); goto cq_next; } wr = aq->q_wr[aq->q.cons].wr; if (wr) { aq->q_wr[aq->q.cons].wr = NULL; list_del_init(&wr->aq_ent); wr->cqe = *cqe; wr->status = atomic_read(&aq->admin_state); complete_all(&wr->work); } ionic_queue_consume_entries(&aq->q, aq->q_wr[aq->q.cons].wqe_strides); cq_next: ionic_queue_produce(&cq->q); cq->color = ionic_color_wrap(cq->q.prod, cq->color); } if (old_prod != cq->q.prod) { ionic_admin_reset_wdog(aq); cq->q.cons = cq->q.prod; ionic_dbell_ring(dev->lif_cfg.dbpage, dev->lif_cfg.cq_qtype, ionic_queue_dbell_val(&cq->q)); queue_work(ionic_evt_workq, &aq->work); } else if (!aq->armed) { aq->armed = true; cq->arm_any_prod = ionic_queue_next(&cq->q, cq->arm_any_prod); ionic_dbell_ring(dev->lif_cfg.dbpage, dev->lif_cfg.cq_qtype, cq->q.dbell | IONIC_CQ_RING_ARM | cq->arm_any_prod); queue_work(ionic_evt_workq, &aq->work); } if (atomic_read(&aq->admin_state) != IONIC_ADMIN_ACTIVE) return; old_prod = aq->q.prod; if (ionic_queue_empty(&aq->q) && !list_empty(&aq->wr_post)) ionic_admin_reset_wdog(aq); if (list_empty(&aq->wr_post)) return; do { u8 *src; int i, src_len; size_t stride_len; wr = list_first_entry(&aq->wr_post, struct ionic_admin_wr, aq_ent); wr_strides = (le16_to_cpu(wr->wqe.len) + ADMIN_WQE_HDR_LEN + (ADMIN_WQE_STRIDE - 1)) >> aq->q.stride_log2; avlbl_strides = ionic_queue_length_remaining(&aq->q); if (wr_strides > avlbl_strides) break; list_move(&wr->aq_ent, &aq->wr_prod); wr->status = aq->q.prod; aq->q_wr[aq->q.prod].wr = wr; aq->q_wr[aq->q.prod].wqe_strides = wr_strides; src_len = le16_to_cpu(wr->wqe.len); src = (uint8_t *)&wr->wqe.cmd; /* First stride */ memcpy(ionic_queue_at_prod(&aq->q), &wr->wqe, ADMIN_WQE_HDR_LEN); stride_len = ADMIN_WQE_STRIDE - ADMIN_WQE_HDR_LEN; if (stride_len > src_len) stride_len = src_len; memcpy(ionic_queue_at_prod(&aq->q) + ADMIN_WQE_HDR_LEN, src, stride_len); ibdev_dbg(&dev->ibdev, "post admin prod %u (%u strides)\n", aq->q.prod, wr_strides); print_hex_dump_debug("wqe ", DUMP_PREFIX_OFFSET, 16, 1, ionic_queue_at_prod(&aq->q), BIT(aq->q.stride_log2), true); ionic_queue_produce(&aq->q); /* Remaining strides */ for (i = stride_len; i < src_len; i += stride_len) { stride_len = ADMIN_WQE_STRIDE; if (i + stride_len > src_len) stride_len = src_len - i; memcpy(ionic_queue_at_prod(&aq->q), src + i, stride_len); print_hex_dump_debug("wqe ", DUMP_PREFIX_OFFSET, 16, 1, ionic_queue_at_prod(&aq->q), BIT(aq->q.stride_log2), true); ionic_queue_produce(&aq->q); } } while (!list_empty(&aq->wr_post)); if (old_prod != aq->q.prod) ionic_dbell_ring(dev->lif_cfg.dbpage, dev->lif_cfg.aq_qtype, ionic_queue_dbell_val(&aq->q)); } static void ionic_admin_dwork(struct work_struct *ws) { struct ionic_ibdev *dev = container_of(ws, struct ionic_ibdev, admin_dwork.work); struct ionic_aq *aq, *bad_aq = NULL; bool do_reschedule = false; unsigned long irqflags; bool do_reset = false; u16 pos; int i; for (i = 0; i < dev->lif_cfg.aq_count; i++) { aq = dev->aq_vec[i]; spin_lock_irqsave(&aq->lock, irqflags); if (ionic_queue_empty(&aq->q)) goto next_aq; /* Reschedule if any queue has outstanding work */ do_reschedule = true; if (time_is_after_eq_jiffies(aq->stamp + IONIC_ADMIN_WARN)) /* Warning threshold not met, nothing to do */ goto next_aq; /* See if polling now makes some progress */ pos = aq->q.cons; ionic_admin_poll_locked(aq); if (pos != aq->q.cons) { ibdev_dbg(&dev->ibdev, "missed event for acq %d\n", aq->cqid); goto next_aq; } if (time_is_after_eq_jiffies(aq->stamp + IONIC_ADMIN_TIMEOUT)) { /* Timeout threshold not met */ ibdev_dbg(&dev->ibdev, "no progress after %ums\n", (u32)jiffies_to_msecs(jiffies - aq->stamp)); goto next_aq; } /* Queue timed out */ bad_aq = aq; do_reset = true; next_aq: spin_unlock_irqrestore(&aq->lock, irqflags); } if (do_reset) /* Reset RDMA lif on a timeout */ ionic_admin_timedout(bad_aq); else if (do_reschedule) /* Try to poll again later */ ionic_admin_reset_dwork(dev); } static void ionic_admin_work(struct work_struct *ws) { struct ionic_aq *aq = container_of(ws, struct ionic_aq, work); unsigned long irqflags; spin_lock_irqsave(&aq->lock, irqflags); ionic_admin_poll_locked(aq); spin_unlock_irqrestore(&aq->lock, irqflags); } static void ionic_admin_post_aq(struct ionic_aq *aq, struct ionic_admin_wr *wr) { unsigned long irqflags; bool poll; wr->status = IONIC_ADMIN_POSTED; wr->aq = aq; spin_lock_irqsave(&aq->lock, irqflags); poll = list_empty(&aq->wr_post); list_add(&wr->aq_ent, &aq->wr_post); if (poll) ionic_admin_poll_locked(aq); spin_unlock_irqrestore(&aq->lock, irqflags); } void ionic_admin_post(struct ionic_ibdev *dev, struct ionic_admin_wr *wr) { int aq_idx; /* Use cpu id for the adminq selection */ aq_idx = raw_smp_processor_id() % dev->lif_cfg.aq_count; ionic_admin_post_aq(dev->aq_vec[aq_idx], wr); } static void ionic_admin_cancel(struct ionic_admin_wr *wr) { struct ionic_aq *aq = wr->aq; unsigned long irqflags; spin_lock_irqsave(&aq->lock, irqflags); if (!list_empty(&wr->aq_ent)) { list_del(&wr->aq_ent); if (wr->status != IONIC_ADMIN_POSTED) aq->q_wr[wr->status].wr = NULL; } spin_unlock_irqrestore(&aq->lock, irqflags); } static int ionic_admin_busy_wait(struct ionic_admin_wr *wr) { struct ionic_aq *aq = wr->aq; unsigned long irqflags; int try_i; for (try_i = 0; try_i < IONIC_ADMIN_BUSY_RETRY_COUNT; ++try_i) { if (completion_done(&wr->work)) return 0; mdelay(IONIC_ADMIN_BUSY_RETRY_MS); spin_lock_irqsave(&aq->lock, irqflags); ionic_admin_poll_locked(aq); spin_unlock_irqrestore(&aq->lock, irqflags); } /* * we timed out. Initiate RDMA LIF reset and indicate * error to caller. */ ionic_admin_timedout(aq); return -ETIMEDOUT; } int ionic_admin_wait(struct ionic_ibdev *dev, struct ionic_admin_wr *wr, enum ionic_admin_flags flags) { int rc, timo; if (flags & IONIC_ADMIN_F_BUSYWAIT) { /* Spin */ rc = ionic_admin_busy_wait(wr); } else if (flags & IONIC_ADMIN_F_INTERRUPT) { /* * Interruptible sleep, 1s timeout * This is used for commands which are safe for the caller * to clean up without killing and resetting the adminq. */ timo = wait_for_completion_interruptible_timeout(&wr->work, HZ); if (timo > 0) rc = 0; else if (timo == 0) rc = -ETIMEDOUT; else rc = timo; } else { /* * Uninterruptible sleep * This is used for commands which are NOT safe for the * caller to clean up. Cleanup must be handled by the * adminq kill and reset process so that host memory is * not corrupted by the device. */ wait_for_completion(&wr->work); rc = 0; } if (rc) { ibdev_warn(&dev->ibdev, "wait status %d\n", rc); ionic_admin_cancel(wr); } else if (wr->status == IONIC_ADMIN_KILLED) { ibdev_dbg(&dev->ibdev, "admin killed\n"); /* No error if admin already killed during teardown */ rc = (flags & IONIC_ADMIN_F_TEARDOWN) ? 0 : -ENODEV; } else if (ionic_v1_cqe_error(&wr->cqe)) { ibdev_warn(&dev->ibdev, "opcode %u error %u\n", wr->wqe.op, be32_to_cpu(wr->cqe.status_length)); rc = -EINVAL; } return rc; } static int ionic_rdma_devcmd(struct ionic_ibdev *dev, struct ionic_admin_ctx *admin) { int rc; rc = ionic_adminq_post_wait(dev->lif_cfg.lif, admin); if (rc) return rc; return ionic_error_to_errno(admin->comp.comp.status); } int ionic_rdma_reset_devcmd(struct ionic_ibdev *dev) { struct ionic_admin_ctx admin = { .work = COMPLETION_INITIALIZER_ONSTACK(admin.work), .cmd.rdma_reset = { .opcode = IONIC_CMD_RDMA_RESET_LIF, .lif_index = cpu_to_le16(dev->lif_cfg.lif_index), }, }; return ionic_rdma_devcmd(dev, &admin); } static int ionic_rdma_queue_devcmd(struct ionic_ibdev *dev, struct ionic_queue *q, u32 qid, u32 cid, u16 opcode) { struct ionic_admin_ctx admin = { .work = COMPLETION_INITIALIZER_ONSTACK(admin.work), .cmd.rdma_queue = { .opcode = opcode, .lif_index = cpu_to_le16(dev->lif_cfg.lif_index), .qid_ver = cpu_to_le32(qid), .cid = cpu_to_le32(cid), .dbid = cpu_to_le16(dev->lif_cfg.dbid), .depth_log2 = q->depth_log2, .stride_log2 = q->stride_log2, .dma_addr = cpu_to_le64(q->dma), }, }; return ionic_rdma_devcmd(dev, &admin); } static void ionic_rdma_admincq_comp(struct ib_cq *ibcq, void *cq_context) { struct ionic_aq *aq = cq_context; unsigned long irqflags; spin_lock_irqsave(&aq->lock, irqflags); aq->armed = false; if (atomic_read(&aq->admin_state) < IONIC_ADMIN_KILLED) queue_work(ionic_evt_workq, &aq->work); spin_unlock_irqrestore(&aq->lock, irqflags); } static void ionic_rdma_admincq_event(struct ib_event *event, void *cq_context) { struct ionic_aq *aq = cq_context; ibdev_err(&aq->dev->ibdev, "admincq event %d\n", event->event); } static struct ionic_vcq *ionic_create_rdma_admincq(struct ionic_ibdev *dev, int comp_vector) { struct ib_cq_init_attr attr = { .cqe = IONIC_AQ_DEPTH, .comp_vector = comp_vector, }; struct ionic_tbl_buf buf = {}; struct ionic_vcq *vcq; struct ionic_cq *cq; int rc; vcq = kzalloc(sizeof(*vcq), GFP_KERNEL); if (!vcq) return ERR_PTR(-ENOMEM); vcq->ibcq.device = &dev->ibdev; vcq->ibcq.comp_handler = ionic_rdma_admincq_comp; vcq->ibcq.event_handler = ionic_rdma_admincq_event; atomic_set(&vcq->ibcq.usecnt, 0); vcq->udma_mask = 1; cq = &vcq->cq[0]; rc = ionic_create_cq_common(vcq, &buf, &attr, NULL, NULL, NULL, NULL, 0); if (rc) goto err_init; rc = ionic_rdma_queue_devcmd(dev, &cq->q, cq->cqid, cq->eqid, IONIC_CMD_RDMA_CREATE_CQ); if (rc) goto err_cmd; return vcq; err_cmd: ionic_destroy_cq_common(dev, cq); err_init: kfree(vcq); return ERR_PTR(rc); } static struct ionic_aq *__ionic_create_rdma_adminq(struct ionic_ibdev *dev, u32 aqid, u32 cqid) { struct ionic_aq *aq; int rc; aq = kzalloc(sizeof(*aq), GFP_KERNEL); if (!aq) return ERR_PTR(-ENOMEM); atomic_set(&aq->admin_state, IONIC_ADMIN_KILLED); aq->dev = dev; aq->aqid = aqid; aq->cqid = cqid; spin_lock_init(&aq->lock); rc = ionic_queue_init(&aq->q, dev->lif_cfg.hwdev, IONIC_EQ_DEPTH, ADMIN_WQE_STRIDE); if (rc) goto err_q; ionic_queue_dbell_init(&aq->q, aq->aqid); aq->q_wr = kcalloc((u32)aq->q.mask + 1, sizeof(*aq->q_wr), GFP_KERNEL); if (!aq->q_wr) { rc = -ENOMEM; goto err_wr; } INIT_LIST_HEAD(&aq->wr_prod); INIT_LIST_HEAD(&aq->wr_post); INIT_WORK(&aq->work, ionic_admin_work); aq->armed = false; return aq; err_wr: ionic_queue_destroy(&aq->q, dev->lif_cfg.hwdev); err_q: kfree(aq); return ERR_PTR(rc); } static void __ionic_destroy_rdma_adminq(struct ionic_ibdev *dev, struct ionic_aq *aq) { kfree(aq->q_wr); ionic_queue_destroy(&aq->q, dev->lif_cfg.hwdev); kfree(aq); } static struct ionic_aq *ionic_create_rdma_adminq(struct ionic_ibdev *dev, u32 aqid, u32 cqid) { struct ionic_aq *aq; int rc; aq = __ionic_create_rdma_adminq(dev, aqid, cqid); if (IS_ERR(aq)) return aq; rc = ionic_rdma_queue_devcmd(dev, &aq->q, aq->aqid, aq->cqid, IONIC_CMD_RDMA_CREATE_ADMINQ); if (rc) goto err_cmd; return aq; err_cmd: __ionic_destroy_rdma_adminq(dev, aq); return ERR_PTR(rc); } static void ionic_flush_qs(struct ionic_ibdev *dev) { struct ionic_qp *qp, *qp_tmp; struct ionic_cq *cq, *cq_tmp; LIST_HEAD(flush_list); unsigned long index; WARN_ON(!irqs_disabled()); /* Flush qp send and recv */ xa_lock(&dev->qp_tbl); xa_for_each(&dev->qp_tbl, index, qp) { kref_get(&qp->qp_kref); list_add_tail(&qp->ibkill_flush_ent, &flush_list); } xa_unlock(&dev->qp_tbl); list_for_each_entry_safe(qp, qp_tmp, &flush_list, ibkill_flush_ent) { ionic_flush_qp(dev, qp); kref_put(&qp->qp_kref, ionic_qp_complete); list_del(&qp->ibkill_flush_ent); } /* Notify completions */ xa_lock(&dev->cq_tbl); xa_for_each(&dev->cq_tbl, index, cq) { kref_get(&cq->cq_kref); list_add_tail(&cq->ibkill_flush_ent, &flush_list); } xa_unlock(&dev->cq_tbl); list_for_each_entry_safe(cq, cq_tmp, &flush_list, ibkill_flush_ent) { ionic_notify_flush_cq(cq); kref_put(&cq->cq_kref, ionic_cq_complete); list_del(&cq->ibkill_flush_ent); } } static void ionic_kill_ibdev(struct ionic_ibdev *dev, bool fatal_path) { unsigned long irqflags; bool do_flush = false; int i; /* Mark AQs for drain and flush the QPs while irq is disabled */ local_irq_save(irqflags); /* Mark the admin queue, flushing at most once */ for (i = 0; i < dev->lif_cfg.aq_count; i++) { struct ionic_aq *aq = dev->aq_vec[i]; spin_lock(&aq->lock); if (atomic_read(&aq->admin_state) != IONIC_ADMIN_KILLED) { atomic_set(&aq->admin_state, IONIC_ADMIN_KILLED); /* Flush incomplete admin commands */ ionic_admin_poll_locked(aq); do_flush = true; } spin_unlock(&aq->lock); } if (do_flush) ionic_flush_qs(dev); local_irq_restore(irqflags); /* Post a fatal event if requested */ if (fatal_path) { struct ib_event ev; ev.device = &dev->ibdev; ev.element.port_num = 1; ev.event = IB_EVENT_DEVICE_FATAL; ib_dispatch_event(&ev); } atomic_set(&dev->admin_state, IONIC_ADMIN_KILLED); } void ionic_kill_rdma_admin(struct ionic_ibdev *dev, bool fatal_path) { enum ionic_admin_state old_state; unsigned long irqflags = 0; int i, rc; if (!dev->aq_vec) return; /* * Admin queues are transitioned from active to paused to killed state. * When in paused state, no new commands are issued to the device, * nor are any completed locally. After resetting the lif, it will be * safe to resume the rdma admin queues in the killed state. Commands * will not be issued to the device, but will complete locally with status * IONIC_ADMIN_KILLED. Handling completion will ensure that creating or * modifying resources fails, but destroying resources succeeds. * If there was a failure resetting the lif using this strategy, * then the state of the device is unknown. */ old_state = atomic_cmpxchg(&dev->admin_state, IONIC_ADMIN_ACTIVE, IONIC_ADMIN_PAUSED); if (old_state != IONIC_ADMIN_ACTIVE) return; /* Pause all the AQs */ local_irq_save(irqflags); for (i = 0; i < dev->lif_cfg.aq_count; i++) { struct ionic_aq *aq = dev->aq_vec[i]; spin_lock(&aq->lock); /* pause rdma admin queues to reset lif */ if (atomic_read(&aq->admin_state) == IONIC_ADMIN_ACTIVE) atomic_set(&aq->admin_state, IONIC_ADMIN_PAUSED); spin_unlock(&aq->lock); } local_irq_restore(irqflags); rc = ionic_rdma_reset_devcmd(dev); if (unlikely(rc)) { ibdev_err(&dev->ibdev, "failed to reset rdma %d\n", rc); ionic_request_rdma_reset(dev->lif_cfg.lif); } ionic_kill_ibdev(dev, fatal_path); } static void ionic_reset_work(struct work_struct *ws) { struct ionic_ibdev *dev = container_of(ws, struct ionic_ibdev, reset_work); ionic_kill_rdma_admin(dev, true); } static bool ionic_next_eqe(struct ionic_eq *eq, struct ionic_v1_eqe *eqe) { struct ionic_v1_eqe *qeqe; bool color; qeqe = ionic_queue_at_prod(&eq->q); color = ionic_v1_eqe_color(qeqe); /* cons is color for eq */ if (eq->q.cons != color) return false; /* Prevent out-of-order reads of the EQE */ dma_rmb(); ibdev_dbg(&eq->dev->ibdev, "poll eq prod %u\n", eq->q.prod); print_hex_dump_debug("eqe ", DUMP_PREFIX_OFFSET, 16, 1, qeqe, BIT(eq->q.stride_log2), true); *eqe = *qeqe; return true; } static void ionic_cq_event(struct ionic_ibdev *dev, u32 cqid, u8 code) { unsigned long irqflags; struct ib_event ibev; struct ionic_cq *cq; xa_lock_irqsave(&dev->cq_tbl, irqflags); cq = xa_load(&dev->cq_tbl, cqid); if (cq) kref_get(&cq->cq_kref); xa_unlock_irqrestore(&dev->cq_tbl, irqflags); if (!cq) { ibdev_dbg(&dev->ibdev, "missing cqid %#x code %u\n", cqid, code); return; } switch (code) { case IONIC_V1_EQE_CQ_NOTIFY: if (cq->vcq->ibcq.comp_handler) cq->vcq->ibcq.comp_handler(&cq->vcq->ibcq, cq->vcq->ibcq.cq_context); break; case IONIC_V1_EQE_CQ_ERR: if (cq->vcq->ibcq.event_handler) { ibev.event = IB_EVENT_CQ_ERR; ibev.device = &dev->ibdev; ibev.element.cq = &cq->vcq->ibcq; cq->vcq->ibcq.event_handler(&ibev, cq->vcq->ibcq.cq_context); } break; default: ibdev_dbg(&dev->ibdev, "unrecognized cqid %#x code %u\n", cqid, code); break; } kref_put(&cq->cq_kref, ionic_cq_complete); } static void ionic_qp_event(struct ionic_ibdev *dev, u32 qpid, u8 code) { unsigned long irqflags; struct ib_event ibev; struct ionic_qp *qp; xa_lock_irqsave(&dev->qp_tbl, irqflags); qp = xa_load(&dev->qp_tbl, qpid); if (qp) kref_get(&qp->qp_kref); xa_unlock_irqrestore(&dev->qp_tbl, irqflags); if (!qp) { ibdev_dbg(&dev->ibdev, "missing qpid %#x code %u\n", qpid, code); return; } ibev.device = &dev->ibdev; ibev.element.qp = &qp->ibqp; switch (code) { case IONIC_V1_EQE_SQ_DRAIN: ibev.event = IB_EVENT_SQ_DRAINED; break; case IONIC_V1_EQE_QP_COMM_EST: ibev.event = IB_EVENT_COMM_EST; break; case IONIC_V1_EQE_QP_LAST_WQE: ibev.event = IB_EVENT_QP_LAST_WQE_REACHED; break; case IONIC_V1_EQE_QP_ERR: ibev.event = IB_EVENT_QP_FATAL; break; case IONIC_V1_EQE_QP_ERR_REQUEST: ibev.event = IB_EVENT_QP_REQ_ERR; break; case IONIC_V1_EQE_QP_ERR_ACCESS: ibev.event = IB_EVENT_QP_ACCESS_ERR; break; default: ibdev_dbg(&dev->ibdev, "unrecognized qpid %#x code %u\n", qpid, code); goto out; } if (qp->ibqp.event_handler) qp->ibqp.event_handler(&ibev, qp->ibqp.qp_context); out: kref_put(&qp->qp_kref, ionic_qp_complete); } static u16 ionic_poll_eq(struct ionic_eq *eq, u16 budget) { struct ionic_ibdev *dev = eq->dev; struct ionic_v1_eqe eqe; u16 npolled = 0; u8 type, code; u32 evt, qid; while (npolled < budget) { if (!ionic_next_eqe(eq, &eqe)) break; ionic_queue_produce(&eq->q); /* cons is color for eq */ eq->q.cons = ionic_color_wrap(eq->q.prod, eq->q.cons); ++npolled; evt = ionic_v1_eqe_evt(&eqe); type = ionic_v1_eqe_evt_type(evt); code = ionic_v1_eqe_evt_code(evt); qid = ionic_v1_eqe_evt_qid(evt); switch (type) { case IONIC_V1_EQE_TYPE_CQ: ionic_cq_event(dev, qid, code); break; case IONIC_V1_EQE_TYPE_QP: ionic_qp_event(dev, qid, code); break; default: ibdev_dbg(&dev->ibdev, "unknown event %#x type %u\n", evt, type); } } return npolled; } static void ionic_poll_eq_work(struct work_struct *work) { struct ionic_eq *eq = container_of(work, struct ionic_eq, work); u32 npolled; if (unlikely(!eq->enable) || WARN_ON(eq->armed)) return; npolled = ionic_poll_eq(eq, IONIC_EQ_WORK_BUDGET); if (npolled == IONIC_EQ_WORK_BUDGET) { ionic_intr_credits(eq->dev->lif_cfg.intr_ctrl, eq->intr, npolled, 0); queue_work(ionic_evt_workq, &eq->work); } else { xchg(&eq->armed, 1); ionic_intr_credits(eq->dev->lif_cfg.intr_ctrl, eq->intr, 0, IONIC_INTR_CRED_UNMASK); } } static irqreturn_t ionic_poll_eq_isr(int irq, void *eqptr) { struct ionic_eq *eq = eqptr; int was_armed; u32 npolled; was_armed = xchg(&eq->armed, 0); if (unlikely(!eq->enable) || !was_armed) return IRQ_HANDLED; npolled = ionic_poll_eq(eq, IONIC_EQ_ISR_BUDGET); if (npolled == IONIC_EQ_ISR_BUDGET) { ionic_intr_credits(eq->dev->lif_cfg.intr_ctrl, eq->intr, npolled, 0); queue_work(ionic_evt_workq, &eq->work); } else { xchg(&eq->armed, 1); ionic_intr_credits(eq->dev->lif_cfg.intr_ctrl, eq->intr, 0, IONIC_INTR_CRED_UNMASK); } return IRQ_HANDLED; } static struct ionic_eq *ionic_create_eq(struct ionic_ibdev *dev, int eqid) { struct ionic_intr_info intr_obj = { }; struct ionic_eq *eq; int rc; eq = kzalloc(sizeof(*eq), GFP_KERNEL); if (!eq) return ERR_PTR(-ENOMEM); eq->dev = dev; rc = ionic_queue_init(&eq->q, dev->lif_cfg.hwdev, IONIC_EQ_DEPTH, sizeof(struct ionic_v1_eqe)); if (rc) goto err_q; eq->eqid = eqid; eq->armed = true; eq->enable = false; INIT_WORK(&eq->work, ionic_poll_eq_work); rc = ionic_intr_alloc(dev->lif_cfg.lif, &intr_obj); if (rc < 0) goto err_intr; eq->irq = intr_obj.vector; eq->intr = intr_obj.index; ionic_queue_dbell_init(&eq->q, eq->eqid); /* cons is color for eq */ eq->q.cons = true; snprintf(eq->name, sizeof(eq->name), "%s-%d-%d-eq", "ionr", dev->lif_cfg.lif_index, eq->eqid); ionic_intr_mask(dev->lif_cfg.intr_ctrl, eq->intr, IONIC_INTR_MASK_SET); ionic_intr_mask_assert(dev->lif_cfg.intr_ctrl, eq->intr, IONIC_INTR_MASK_SET); ionic_intr_coal_init(dev->lif_cfg.intr_ctrl, eq->intr, 0); ionic_intr_clean(dev->lif_cfg.intr_ctrl, eq->intr); eq->enable = true; rc = request_irq(eq->irq, ionic_poll_eq_isr, 0, eq->name, eq); if (rc) goto err_irq; rc = ionic_rdma_queue_devcmd(dev, &eq->q, eq->eqid, eq->intr, IONIC_CMD_RDMA_CREATE_EQ); if (rc) goto err_cmd; ionic_intr_mask(dev->lif_cfg.intr_ctrl, eq->intr, IONIC_INTR_MASK_CLEAR); return eq; err_cmd: eq->enable = false; free_irq(eq->irq, eq); flush_work(&eq->work); err_irq: ionic_intr_free(dev->lif_cfg.lif, eq->intr); err_intr: ionic_queue_destroy(&eq->q, dev->lif_cfg.hwdev); err_q: kfree(eq); return ERR_PTR(rc); } static void ionic_destroy_eq(struct ionic_eq *eq) { struct ionic_ibdev *dev = eq->dev; eq->enable = false; free_irq(eq->irq, eq); flush_work(&eq->work); ionic_intr_free(dev->lif_cfg.lif, eq->intr); ionic_queue_destroy(&eq->q, dev->lif_cfg.hwdev); kfree(eq); } int ionic_create_rdma_admin(struct ionic_ibdev *dev) { int eq_i = 0, aq_i = 0, rc = 0; struct ionic_vcq *vcq; struct ionic_aq *aq; struct ionic_eq *eq; dev->eq_vec = NULL; dev->aq_vec = NULL; INIT_WORK(&dev->reset_work, ionic_reset_work); INIT_DELAYED_WORK(&dev->admin_dwork, ionic_admin_dwork); atomic_set(&dev->admin_state, IONIC_ADMIN_KILLED); if (dev->lif_cfg.aq_count > IONIC_AQ_COUNT) { ibdev_dbg(&dev->ibdev, "limiting adminq count to %d\n", IONIC_AQ_COUNT); dev->lif_cfg.aq_count = IONIC_AQ_COUNT; } if (dev->lif_cfg.eq_count > IONIC_EQ_COUNT) { dev_dbg(&dev->ibdev.dev, "limiting eventq count to %d\n", IONIC_EQ_COUNT); dev->lif_cfg.eq_count = IONIC_EQ_COUNT; } /* need at least two eq and one aq */ if (dev->lif_cfg.eq_count < IONIC_EQ_COUNT_MIN || dev->lif_cfg.aq_count < IONIC_AQ_COUNT_MIN) { rc = -EINVAL; goto out; } dev->eq_vec = kmalloc_array(dev->lif_cfg.eq_count, sizeof(*dev->eq_vec), GFP_KERNEL); if (!dev->eq_vec) { rc = -ENOMEM; goto out; } for (eq_i = 0; eq_i < dev->lif_cfg.eq_count; ++eq_i) { eq = ionic_create_eq(dev, eq_i + dev->lif_cfg.eq_base); if (IS_ERR(eq)) { rc = PTR_ERR(eq); if (eq_i < IONIC_EQ_COUNT_MIN) { ibdev_err(&dev->ibdev, "fail create eq %pe\n", eq); goto out; } /* ok, just fewer eq than device supports */ ibdev_dbg(&dev->ibdev, "eq count %d want %d rc %pe\n", eq_i, dev->lif_cfg.eq_count, eq); rc = 0; break; } dev->eq_vec[eq_i] = eq; } dev->lif_cfg.eq_count = eq_i; dev->aq_vec = kmalloc_array(dev->lif_cfg.aq_count, sizeof(*dev->aq_vec), GFP_KERNEL); if (!dev->aq_vec) { rc = -ENOMEM; goto out; } /* Create one CQ per AQ */ for (aq_i = 0; aq_i < dev->lif_cfg.aq_count; ++aq_i) { vcq = ionic_create_rdma_admincq(dev, aq_i % eq_i); if (IS_ERR(vcq)) { rc = PTR_ERR(vcq); if (!aq_i) { ibdev_err(&dev->ibdev, "failed to create acq %pe\n", vcq); goto out; } /* ok, just fewer adminq than device supports */ ibdev_dbg(&dev->ibdev, "acq count %d want %d rc %pe\n", aq_i, dev->lif_cfg.aq_count, vcq); break; } aq = ionic_create_rdma_adminq(dev, aq_i + dev->lif_cfg.aq_base, vcq->cq[0].cqid); if (IS_ERR(aq)) { /* Clean up the dangling CQ */ ionic_destroy_cq_common(dev, &vcq->cq[0]); kfree(vcq); rc = PTR_ERR(aq); if (!aq_i) { ibdev_err(&dev->ibdev, "failed to create aq %pe\n", aq); goto out; } /* ok, just fewer adminq than device supports */ ibdev_dbg(&dev->ibdev, "aq count %d want %d rc %pe\n", aq_i, dev->lif_cfg.aq_count, aq); break; } vcq->ibcq.cq_context = aq; aq->vcq = vcq; atomic_set(&aq->admin_state, IONIC_ADMIN_ACTIVE); dev->aq_vec[aq_i] = aq; } atomic_set(&dev->admin_state, IONIC_ADMIN_ACTIVE); out: dev->lif_cfg.eq_count = eq_i; dev->lif_cfg.aq_count = aq_i; return rc; } void ionic_destroy_rdma_admin(struct ionic_ibdev *dev) { struct ionic_vcq *vcq; struct ionic_aq *aq; struct ionic_eq *eq; /* * Killing the admin before destroy makes sure all admin and * completions are flushed. admin_state = IONIC_ADMIN_KILLED * stops queueing up further works. */ cancel_delayed_work_sync(&dev->admin_dwork); cancel_work_sync(&dev->reset_work); if (dev->aq_vec) { while (dev->lif_cfg.aq_count > 0) { aq = dev->aq_vec[--dev->lif_cfg.aq_count]; vcq = aq->vcq; cancel_work_sync(&aq->work); __ionic_destroy_rdma_adminq(dev, aq); if (vcq) { ionic_destroy_cq_common(dev, &vcq->cq[0]); kfree(vcq); } } kfree(dev->aq_vec); } if (dev->eq_vec) { while (dev->lif_cfg.eq_count > 0) { eq = dev->eq_vec[--dev->lif_cfg.eq_count]; ionic_destroy_eq(eq); } kfree(dev->eq_vec); } }