123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297 |
- /* SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB */
- /* Copyright (c) 2015 - 2021 Intel Corporation */
- #ifndef IRDMA_VERBS_H
- #define IRDMA_VERBS_H
- #define IRDMA_MAX_SAVED_PHY_PGADDR 4
- #define IRDMA_FLUSH_DELAY_MS 20
- #define IRDMA_PKEY_TBL_SZ 1
- #define IRDMA_DEFAULT_PKEY 0xFFFF
- struct irdma_ucontext {
- struct ib_ucontext ibucontext;
- struct irdma_device *iwdev;
- struct rdma_user_mmap_entry *db_mmap_entry;
- struct list_head cq_reg_mem_list;
- spinlock_t cq_reg_mem_list_lock; /* protect CQ memory list */
- struct list_head qp_reg_mem_list;
- spinlock_t qp_reg_mem_list_lock; /* protect QP memory list */
- int abi_ver;
- bool legacy_mode;
- };
- struct irdma_pd {
- struct ib_pd ibpd;
- struct irdma_sc_pd sc_pd;
- };
- union irdma_sockaddr {
- struct sockaddr_in saddr_in;
- struct sockaddr_in6 saddr_in6;
- };
- struct irdma_av {
- u8 macaddr[16];
- struct rdma_ah_attr attrs;
- union irdma_sockaddr sgid_addr;
- union irdma_sockaddr dgid_addr;
- u8 net_type;
- };
- struct irdma_ah {
- struct ib_ah ibah;
- struct irdma_sc_ah sc_ah;
- struct irdma_pd *pd;
- struct irdma_av av;
- u8 sgid_index;
- union ib_gid dgid;
- struct hlist_node list;
- refcount_t refcnt;
- struct irdma_ah *parent_ah; /* AH from cached list */
- };
- struct irdma_hmc_pble {
- union {
- u32 idx;
- dma_addr_t addr;
- };
- };
- struct irdma_cq_mr {
- struct irdma_hmc_pble cq_pbl;
- dma_addr_t shadow;
- bool split;
- };
- struct irdma_qp_mr {
- struct irdma_hmc_pble sq_pbl;
- struct irdma_hmc_pble rq_pbl;
- dma_addr_t shadow;
- struct page *sq_page;
- };
- struct irdma_cq_buf {
- struct irdma_dma_mem kmem_buf;
- struct irdma_cq_uk cq_uk;
- struct irdma_hw *hw;
- struct list_head list;
- struct work_struct work;
- };
- struct irdma_pbl {
- struct list_head list;
- union {
- struct irdma_qp_mr qp_mr;
- struct irdma_cq_mr cq_mr;
- };
- bool pbl_allocated:1;
- bool on_list:1;
- u64 user_base;
- struct irdma_pble_alloc pble_alloc;
- struct irdma_mr *iwmr;
- };
- struct irdma_mr {
- union {
- struct ib_mr ibmr;
- struct ib_mw ibmw;
- };
- struct ib_umem *region;
- u16 type;
- u32 page_cnt;
- u64 page_size;
- u32 npages;
- u32 stag;
- u64 len;
- u64 pgaddrmem[IRDMA_MAX_SAVED_PHY_PGADDR];
- struct irdma_pbl iwpbl;
- };
- struct irdma_cq {
- struct ib_cq ibcq;
- struct irdma_sc_cq sc_cq;
- u16 cq_head;
- u16 cq_size;
- u16 cq_num;
- bool user_mode;
- atomic_t armed;
- enum irdma_cmpl_notify last_notify;
- u32 polled_cmpls;
- u32 cq_mem_size;
- struct irdma_dma_mem kmem;
- struct irdma_dma_mem kmem_shadow;
- spinlock_t lock; /* for poll cq */
- struct irdma_pbl *iwpbl;
- struct irdma_pbl *iwpbl_shadow;
- struct list_head resize_list;
- struct irdma_cq_poll_info cur_cqe;
- struct list_head cmpl_generated;
- };
- struct irdma_cmpl_gen {
- struct list_head list;
- struct irdma_cq_poll_info cpi;
- };
- struct disconn_work {
- struct work_struct work;
- struct irdma_qp *iwqp;
- };
- struct iw_cm_id;
- struct irdma_qp_kmode {
- struct irdma_dma_mem dma_mem;
- struct irdma_sq_uk_wr_trk_info *sq_wrid_mem;
- u64 *rq_wrid_mem;
- };
- struct irdma_qp {
- struct ib_qp ibqp;
- struct irdma_sc_qp sc_qp;
- struct irdma_device *iwdev;
- struct irdma_cq *iwscq;
- struct irdma_cq *iwrcq;
- struct irdma_pd *iwpd;
- struct rdma_user_mmap_entry *push_wqe_mmap_entry;
- struct rdma_user_mmap_entry *push_db_mmap_entry;
- struct irdma_qp_host_ctx_info ctx_info;
- union {
- struct irdma_iwarp_offload_info iwarp_info;
- struct irdma_roce_offload_info roce_info;
- };
- union {
- struct irdma_tcp_offload_info tcp_info;
- struct irdma_udp_offload_info udp_info;
- };
- struct irdma_ah roce_ah;
- struct list_head teardown_entry;
- refcount_t refcnt;
- struct iw_cm_id *cm_id;
- struct irdma_cm_node *cm_node;
- struct delayed_work dwork_flush;
- struct ib_mr *lsmm_mr;
- atomic_t hw_mod_qp_pend;
- enum ib_qp_state ibqp_state;
- u32 qp_mem_size;
- u32 last_aeq;
- int max_send_wr;
- int max_recv_wr;
- atomic_t close_timer_started;
- spinlock_t lock; /* serialize posting WRs to SQ/RQ */
- struct irdma_qp_context *iwqp_context;
- void *pbl_vbase;
- dma_addr_t pbl_pbase;
- struct page *page;
- u8 active_conn : 1;
- u8 user_mode : 1;
- u8 hte_added : 1;
- u8 flush_issued : 1;
- u8 sig_all : 1;
- u8 pau_mode : 1;
- u8 suspend_pending : 1;
- u8 rsvd : 1;
- u8 iwarp_state;
- u16 term_sq_flush_code;
- u16 term_rq_flush_code;
- u8 hw_iwarp_state;
- u8 hw_tcp_state;
- struct irdma_qp_kmode kqp;
- struct irdma_dma_mem host_ctx;
- struct timer_list terminate_timer;
- struct irdma_pbl *iwpbl;
- struct irdma_dma_mem q2_ctx_mem;
- struct irdma_dma_mem ietf_mem;
- struct completion free_qp;
- wait_queue_head_t waitq;
- wait_queue_head_t mod_qp_waitq;
- u8 rts_ae_rcvd;
- };
- enum irdma_mmap_flag {
- IRDMA_MMAP_IO_NC,
- IRDMA_MMAP_IO_WC,
- };
- struct irdma_user_mmap_entry {
- struct rdma_user_mmap_entry rdma_entry;
- u64 bar_offset;
- u8 mmap_flag;
- };
- static inline u16 irdma_fw_major_ver(struct irdma_sc_dev *dev)
- {
- return (u16)FIELD_GET(IRDMA_FW_VER_MAJOR, dev->feature_info[IRDMA_FEATURE_FW_INFO]);
- }
- static inline u16 irdma_fw_minor_ver(struct irdma_sc_dev *dev)
- {
- return (u16)FIELD_GET(IRDMA_FW_VER_MINOR, dev->feature_info[IRDMA_FEATURE_FW_INFO]);
- }
- static inline void set_ib_wc_op_sq(struct irdma_cq_poll_info *cq_poll_info,
- struct ib_wc *entry)
- {
- switch (cq_poll_info->op_type) {
- case IRDMA_OP_TYPE_RDMA_WRITE:
- case IRDMA_OP_TYPE_RDMA_WRITE_SOL:
- entry->opcode = IB_WC_RDMA_WRITE;
- break;
- case IRDMA_OP_TYPE_RDMA_READ_INV_STAG:
- case IRDMA_OP_TYPE_RDMA_READ:
- entry->opcode = IB_WC_RDMA_READ;
- break;
- case IRDMA_OP_TYPE_SEND_SOL:
- case IRDMA_OP_TYPE_SEND_SOL_INV:
- case IRDMA_OP_TYPE_SEND_INV:
- case IRDMA_OP_TYPE_SEND:
- entry->opcode = IB_WC_SEND;
- break;
- case IRDMA_OP_TYPE_FAST_REG_NSMR:
- entry->opcode = IB_WC_REG_MR;
- break;
- case IRDMA_OP_TYPE_INV_STAG:
- entry->opcode = IB_WC_LOCAL_INV;
- break;
- default:
- entry->status = IB_WC_GENERAL_ERR;
- }
- }
- static inline void set_ib_wc_op_rq(struct irdma_cq_poll_info *cq_poll_info,
- struct ib_wc *entry, bool send_imm_support)
- {
- /**
- * iWARP does not support sendImm, so the presence of Imm data
- * must be WriteImm.
- */
- if (!send_imm_support) {
- entry->opcode = cq_poll_info->imm_valid ?
- IB_WC_RECV_RDMA_WITH_IMM :
- IB_WC_RECV;
- return;
- }
- switch (cq_poll_info->op_type) {
- case IB_OPCODE_RDMA_WRITE_ONLY_WITH_IMMEDIATE:
- case IB_OPCODE_RDMA_WRITE_LAST_WITH_IMMEDIATE:
- entry->opcode = IB_WC_RECV_RDMA_WITH_IMM;
- break;
- default:
- entry->opcode = IB_WC_RECV;
- }
- }
- void irdma_mcast_mac(u32 *ip_addr, u8 *mac, bool ipv4);
- int irdma_ib_register_device(struct irdma_device *iwdev);
- void irdma_ib_unregister_device(struct irdma_device *iwdev);
- void irdma_ib_dealloc_device(struct ib_device *ibdev);
- void irdma_ib_qp_event(struct irdma_qp *iwqp, enum irdma_qp_event_type event);
- void irdma_generate_flush_completions(struct irdma_qp *iwqp);
- void irdma_remove_cmpls_list(struct irdma_cq *iwcq);
- int irdma_generated_cmpls(struct irdma_cq *iwcq, struct irdma_cq_poll_info *cq_poll_info);
- #endif /* IRDMA_VERBS_H */
|