/freebsd/sys/dev/mlx4/mlx4_ib/ |
H A D | mlx4_ib_main.c | 1225 ge->port = mqp->port; in add_gid_entry() 1258 if (!mqp->port) in mlx4_ib_add_mc() 1804 err = mlx4_multicast_attach(mdev->dev, &mqp->mqp, gid->raw, mqp->port, in mlx4_ib_mcg_attach() 1805 !!(mqp->flags & in mlx4_ib_mcg_attach() 1815 err = mlx4_multicast_attach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_attach() 1838 mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_attach() 1841 mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_attach() 1896 err = mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_detach() 1902 err = mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_detach() 2962 if (mqp->sq.tail != mqp->sq.head) { in mlx4_ib_handle_catas_error() [all …]
|
H A D | mlx4_ib_qp.c | 874 qp->mqp.qpn |= (1 << 23); in create_qp_common() 1035 qp->mqp.qpn); in destroy_qp_common() 1288 if (is_qp0(dev, mqp)) in _mlx4_ib_destroy_qp() 1291 if (dev->qp1_proxy[mqp->port - 1] == mqp) { in _mlx4_ib_destroy_qp() 1297 if (mqp->counter_index) in _mlx4_ib_destroy_qp() 1302 if (is_sqp(dev, mqp)) in _mlx4_ib_destroy_qp() 1303 kfree(to_msqp(mqp)); in _mlx4_ib_destroy_qp() 1305 kfree(mqp); in _mlx4_ib_destroy_qp() 1514 struct mlx4_ib_qp *mqp, in mlx4_set_path() argument 1527 struct mlx4_ib_qp *mqp, in mlx4_set_alt_path() argument [all …]
|
H A D | mlx4_ib.h | 295 struct mlx4_qp mqp; member 684 static inline struct mlx4_ib_qp *to_mibqp(struct mlx4_qp *mqp) in to_mibqp() argument 686 return container_of(mqp, struct mlx4_ib_qp, mqp); in to_mibqp() 820 int mlx4_ib_add_mc(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp, 878 int mlx4_ib_steer_qp_reg(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp,
|
H A D | mlx4_ib_cq.c | 654 struct mlx4_qp *mqp; in mlx4_ib_poll_one() local 702 (be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK) != (*cur_qp)->mqp.qpn) { in mlx4_ib_poll_one() 708 mqp = __mlx4_qp_lookup(to_mdev(cq->ibcq.device)->dev, in mlx4_ib_poll_one() 710 *cur_qp = to_mibqp(mqp); in mlx4_ib_poll_one()
|
/freebsd/contrib/ofed/libmlx5/ |
H A D | mlx5.c | 649 struct mlx5_qp *mqp = to_mqp(qp_in); in mlx5dv_get_qp() local 652 qp_out->dbrec = mqp->db; in mlx5dv_get_qp() 654 if (mqp->sq_buf_size) in mlx5dv_get_qp() 658 qp_out->sq.buf = (void *)((uintptr_t)mqp->buf.buf + mqp->sq.offset); in mlx5dv_get_qp() 659 qp_out->sq.wqe_cnt = mqp->sq.wqe_cnt; in mlx5dv_get_qp() 660 qp_out->sq.stride = 1 << mqp->sq.wqe_shift; in mlx5dv_get_qp() 662 qp_out->rq.buf = (void *)((uintptr_t)mqp->buf.buf + mqp->rq.offset); in mlx5dv_get_qp() 663 qp_out->rq.wqe_cnt = mqp->rq.wqe_cnt; in mlx5dv_get_qp() 666 qp_out->bf.reg = mqp->bf->reg; in mlx5dv_get_qp() 668 if (mqp->bf->uuarn > 0) in mlx5dv_get_qp() [all …]
|
H A D | cq.c | 369 struct mlx5_qp *mqp; in get_resp_ctx_v1() local 379 mqp = rsc_to_mqp(*cur_rsc); in get_resp_ctx_v1() 380 if (mqp->verbs_qp.qp.srq) { in get_resp_ctx_v1() 532 struct mlx5_qp *mqp; in mlx5_parse_cqe() local 550 mqp = get_req_context(mctx, cur_rsc, in mlx5_parse_cqe() 553 if (unlikely(!mqp)) in mlx5_parse_cqe() 555 wq = &mqp->sq; in mlx5_parse_cqe() 576 mqp, wqe_ctr, cqe, wc_byte_len); in mlx5_parse_cqe() 649 mqp = get_req_context(mctx, cur_rsc, in mlx5_parse_cqe() 651 if (unlikely(!mqp)) in mlx5_parse_cqe() [all …]
|
H A D | verbs.c | 1579 struct mlx5_qp *mqp = to_mqp(qp); in mlx5_modify_qp() local 1584 if (mqp->rss_qp) in mlx5_modify_qp() 1594 mqp->qp_cap_cache |= in mlx5_modify_qp() 1601 mqp->max_tso = in mlx5_modify_qp() 1624 mlx5_cq_clean(to_mcq(qp->recv_cq), mqp->rsc.rsn, in mlx5_modify_qp() 1631 mlx5_init_qp_indices(mqp); in mlx5_modify_qp() 1632 db = mqp->db; in mlx5_modify_qp() 1649 mlx5_spin_lock(&mqp->rq.lock); in mlx5_modify_qp() 1650 mqp->db[MLX5_RCV_DBR] = htobe32(mqp->rq.head & 0xffff); in mlx5_modify_qp() 1651 mlx5_spin_unlock(&mqp->rq.lock); in mlx5_modify_qp()
|
/freebsd/sys/dev/mlx5/mlx5_fpga/ |
H A D | mlx5fpga_conn.c | 166 ctrl->qpn_ds = cpu_to_be32(size | (conn->qp.mqp.qpn << 8)); in mlx5_fpga_conn_post_send() 376 conn = container_of(mqp, struct mlx5_fpga_conn, qp.mqp); in mlx5_fpga_conn_event() 595 err = mlx5_core_create_qp(mdev, &conn->qp.mqp, in, inlen); in mlx5_fpga_conn_create_qp() 599 conn->qp.mqp.event = mlx5_fpga_conn_event; in mlx5_fpga_conn_create_qp() 600 mlx5_fpga_dbg(fdev, "Created QP #0x%x\n", conn->qp.mqp.qpn); in mlx5_fpga_conn_create_qp() 653 mlx5_core_destroy_qp(conn->fdev->mdev, &conn->qp.mqp); in mlx5_fpga_conn_destroy_qp() 668 &conn->qp.mqp); in mlx5_fpga_conn_reset_qp() 696 &conn->qp.mqp); in mlx5_fpga_conn_init_qp() 741 &conn->qp.mqp); in mlx5_fpga_conn_rtr_qp() 778 &conn->qp.mqp); in mlx5_fpga_conn_rts_qp() [all …]
|
H A D | conn.h | 67 struct mlx5_core_qp mqp; member
|
/freebsd/sys/dev/mlx5/mlx5_ib/ |
H A D | mlx5_ib_qp.c | 1309 qp->trans_qp.base.mqp.qpn = qp->sq.wqe_cnt ? sq->base.mqp.qpn : in create_raw_packet_qp() 1310 rq->base.mqp.qpn; in create_raw_packet_qp() 1530 qp->trans_qp.base.mqp.qpn = 0; in create_rss_raw_qp_tir() 1857 base->mqp.event = mlx5_ib_qp_event; in create_qp_common() 2012 NULL, &base->mqp); in destroy_qp_common() 2022 base->mqp.qpn); in destroy_qp_common() 2054 base->mqp.qpn); in destroy_qp_common() 2193 struct mlx5_ib_qp *mqp = to_mqp(qp); in mlx5_ib_destroy_qp() local 2198 destroy_qp_common(dev, mqp, udata); in mlx5_ib_destroy_qp() 2200 kfree(mqp); in mlx5_ib_destroy_qp() [all …]
|
H A D | mlx5_ib_main.c | 2369 struct mlx5_ib_qp *mqp = to_mqp(qp); in mlx5_ib_create_flow() local 2441 if (mqp->flags & MLX5_IB_QP_RSS) in mlx5_ib_create_flow() 2442 dst->tir_num = mqp->rss_qp.tirn; in mlx5_ib_create_flow() 2601 struct mlx5_ib_qp *mqp; in mlx5_ib_handle_internal_error() local 2615 if (mqp->sq.tail != mqp->sq.head) { in mlx5_ib_handle_internal_error() 2616 send_mcq = to_mcq(mqp->ibqp.send_cq); in mlx5_ib_handle_internal_error() 2619 mqp->ibqp.send_cq->comp_handler) { in mlx5_ib_handle_internal_error() 2631 if (!mqp->ibqp.srq) { in mlx5_ib_handle_internal_error() 2632 if (mqp->rq.tail != mqp->rq.head) { in mlx5_ib_handle_internal_error() 2633 recv_mcq = to_mcq(mqp->ibqp.recv_cq); in mlx5_ib_handle_internal_error() [all …]
|
H A D | mlx5_ib.h | 312 struct mlx5_core_qp mqp; member 823 static inline struct mlx5_ib_qp *to_mibqp(struct mlx5_core_qp *mqp) in to_mibqp() argument 825 return container_of(mqp, struct mlx5_ib_qp_base, mqp)->container_mibqp; in to_mibqp()
|
H A D | mlx5_ib_cq.c | 521 struct mlx5_core_qp *mqp; in mlx5_poll_one() local 566 mqp = __mlx5_qp_lookup(dev->mdev, qpn); in mlx5_poll_one() 567 *cur_qp = to_mibqp(mqp); in mlx5_poll_one()
|
H A D | mlx5_ib_devx.c | 654 rq->base.mqp.qpn) == obj_id || in devx_is_valid_obj_id() 656 sq->base.mqp.qpn) == obj_id || in devx_is_valid_obj_id()
|
/freebsd/contrib/ofed/libmlx4/ |
H A D | verbs.c | 995 struct mlx4_qp *mqp = to_mqp(qp); in mlx4_modify_qp() local 1005 mqp->link_layer = port_attr.link_layer; in mlx4_modify_qp() 1013 if ((mqp->link_layer == IBV_LINK_LAYER_INFINIBAND) && in mlx4_modify_qp() 1015 mqp->qp_cap_cache |= MLX4_CSUM_SUPPORT_UD_OVER_IB | in mlx4_modify_qp() 1019 if ((mqp->link_layer == IBV_LINK_LAYER_ETHERNET) && in mlx4_modify_qp() 1021 mqp->qp_cap_cache |= MLX4_CSUM_SUPPORT_RAW_OVER_ETH | in mlx4_modify_qp()
|