/* QLogic qed NIC Driver * Copyright (c) 2015 QLogic Corporation * * This software is available under the terms of the GNU General Public License * (GPL) Version 2, available from the file COPYING in the main directory of * this source tree. */ #ifndef _QED_SP_H #define _QED_SP_H #include #include #include #include #include #include #include "qed.h" #include "qed_hsi.h" enum spq_mode { QED_SPQ_MODE_BLOCK, /* Client will poll a designated mem. address */ QED_SPQ_MODE_CB, /* Client supplies a callback */ QED_SPQ_MODE_EBLOCK, /* QED should block until completion */ }; struct qed_spq_comp_cb { void (*function)(struct qed_hwfn *, void *, union event_ring_data *, u8 fw_return_code); void *cookie; }; /** * @brief qed_eth_cqe_completion - handles the completion of a * ramrod on the cqe ring * * @param p_hwfn * @param cqe * * @return int */ int qed_eth_cqe_completion(struct qed_hwfn *p_hwfn, struct eth_slow_path_rx_cqe *cqe); /** * @file * * QED Slow-hwfn queue interface */ union ramrod_data { struct pf_start_ramrod_data pf_start; struct pf_update_ramrod_data pf_update; struct rx_queue_start_ramrod_data rx_queue_start; struct rx_queue_update_ramrod_data rx_queue_update; struct rx_queue_stop_ramrod_data rx_queue_stop; struct tx_queue_start_ramrod_data tx_queue_start; struct tx_queue_stop_ramrod_data tx_queue_stop; struct vport_start_ramrod_data vport_start; struct vport_stop_ramrod_data vport_stop; struct vport_update_ramrod_data vport_update; struct core_rx_start_ramrod_data core_rx_queue_start; struct core_rx_stop_ramrod_data core_rx_queue_stop; struct core_tx_start_ramrod_data core_tx_queue_start; struct core_tx_stop_ramrod_data core_tx_queue_stop; struct vport_filter_update_ramrod_data vport_filter_update; struct rdma_init_func_ramrod_data rdma_init_func; struct rdma_close_func_ramrod_data rdma_close_func; struct rdma_register_tid_ramrod_data rdma_register_tid; struct rdma_deregister_tid_ramrod_data rdma_deregister_tid; struct roce_create_qp_resp_ramrod_data roce_create_qp_resp; struct roce_create_qp_req_ramrod_data roce_create_qp_req; struct roce_modify_qp_resp_ramrod_data roce_modify_qp_resp; struct roce_modify_qp_req_ramrod_data roce_modify_qp_req; struct roce_query_qp_resp_ramrod_data roce_query_qp_resp; struct roce_query_qp_req_ramrod_data roce_query_qp_req; struct roce_destroy_qp_resp_ramrod_data roce_destroy_qp_resp; struct roce_destroy_qp_req_ramrod_data roce_destroy_qp_req; struct rdma_create_cq_ramrod_data rdma_create_cq; struct rdma_destroy_cq_ramrod_data rdma_destroy_cq; struct rdma_srq_create_ramrod_data rdma_create_srq; struct rdma_srq_destroy_ramrod_data rdma_destroy_srq; struct rdma_srq_modify_ramrod_data rdma_modify_srq; struct roce_init_func_ramrod_data roce_init_func; struct iscsi_slow_path_hdr iscsi_empty; struct iscsi_init_ramrod_params iscsi_init; struct iscsi_spe_func_dstry iscsi_destroy; struct iscsi_spe_conn_offload iscsi_conn_offload; struct iscsi_conn_update_ramrod_params iscsi_conn_update; struct iscsi_spe_conn_termination iscsi_conn_terminate; struct vf_start_ramrod_data vf_start; struct vf_stop_ramrod_data vf_stop; }; #define EQ_MAX_CREDIT 0xffffffff enum spq_priority { QED_SPQ_PRIORITY_NORMAL, QED_SPQ_PRIORITY_HIGH, }; union qed_spq_req_comp { struct qed_spq_comp_cb cb; u64 *done_addr; }; struct qed_spq_comp_done { u64 done; u8 fw_return_code; }; struct qed_spq_entry { struct list_head list; u8 flags; /* HSI slow path element */ struct slow_path_element elem; union ramrod_data ramrod; enum spq_priority priority; /* pending queue for this entry */ struct list_head *queue; enum spq_mode comp_mode; struct qed_spq_comp_cb comp_cb; struct qed_spq_comp_done comp_done; /* SPQ_MODE_EBLOCK */ }; struct qed_eq { struct qed_chain chain; u8 eq_sb_index; /* index within the SB */ __le16 *p_fw_cons; /* ptr to index value */ }; struct qed_consq { struct qed_chain chain; }; struct qed_spq { spinlock_t lock; /* SPQ lock */ struct list_head unlimited_pending; struct list_head pending; struct list_head completion_pending; struct list_head free_pool; struct qed_chain chain; /* allocated dma-able memory for spq entries (+ramrod data) */ dma_addr_t p_phys; struct qed_spq_entry *p_virt; #define SPQ_RING_SIZE \ (CORE_SPQE_PAGE_SIZE_BYTES / sizeof(struct slow_path_element)) /* Bitmap for handling out-of-order completions */ DECLARE_BITMAP(p_comp_bitmap, SPQ_RING_SIZE); u8 comp_bitmap_idx; /* Statistics */ u32 unlimited_pending_count; u32 normal_count; u32 high_count; u32 comp_sent_count; u32 comp_count; u32 cid; }; /** * @brief qed_spq_post - Posts a Slow hwfn request to FW, or lacking that * Pends it to the future list. * * @param p_hwfn * @param p_req * * @return int */ int qed_spq_post(struct qed_hwfn *p_hwfn, struct qed_spq_entry *p_ent, u8 *fw_return_code); /** * @brief qed_spq_allocate - Alloocates & initializes the SPQ and EQ. * * @param p_hwfn * * @return int */ int qed_spq_alloc(struct qed_hwfn *p_hwfn); /** * @brief qed_spq_setup - Reset the SPQ to its start state. * * @param p_hwfn */ void qed_spq_setup(struct qed_hwfn *p_hwfn); /** * @brief qed_spq_deallocate - Deallocates the given SPQ struct. * * @param p_hwfn */ void qed_spq_free(struct qed_hwfn *p_hwfn); /** * @brief qed_spq_get_entry - Obtain an entrry from the spq * free pool list. * * * * @param p_hwfn * @param pp_ent * * @return int */ int qed_spq_get_entry(struct qed_hwfn *p_hwfn, struct qed_spq_entry **pp_ent); /** * @brief qed_spq_return_entry - Return an entry to spq free * pool list * * @param p_hwfn * @param p_ent */ void qed_spq_return_entry(struct qed_hwfn *p_hwfn, struct qed_spq_entry *p_ent); /** * @brief qed_eq_allocate - Allocates & initializes an EQ struct * * @param p_hwfn * @param num_elem number of elements in the eq * * @return struct qed_eq* - a newly allocated structure; NULL upon error. */ struct qed_eq *qed_eq_alloc(struct qed_hwfn *p_hwfn, u16 num_elem); /** * @brief qed_eq_setup - Reset the SPQ to its start state. * * @param p_hwfn * @param p_eq */ void qed_eq_setup(struct qed_hwfn *p_hwfn, struct qed_eq *p_eq); /** * @brief qed_eq_deallocate - deallocates the given EQ struct. * * @param p_hwfn * @param p_eq */ void qed_eq_free(struct qed_hwfn *p_hwfn, struct qed_eq *p_eq); /** * @brief qed_eq_prod_update - update the FW with default EQ producer * * @param p_hwfn * @param prod */ void qed_eq_prod_update(struct qed_hwfn *p_hwfn, u16 prod); /** * @brief qed_eq_completion - Completes currently pending EQ elements * * @param p_hwfn * @param cookie * * @return int */ int qed_eq_completion(struct qed_hwfn *p_hwfn, void *cookie); /** * @brief qed_spq_completion - Completes a single event * * @param p_hwfn * @param echo - echo value from cookie (used for determining completion) * @param p_data - data from cookie (used in callback function if applicable) * * @return int */ int qed_spq_completion(struct qed_hwfn *p_hwfn, __le16 echo, u8 fw_return_code, union event_ring_data *p_data); /** * @brief qed_spq_get_cid - Given p_hwfn, return cid for the hwfn's SPQ * * @param p_hwfn * * @return u32 - SPQ CID */ u32 qed_spq_get_cid(struct qed_hwfn *p_hwfn); /** * @brief qed_consq_alloc - Allocates & initializes an ConsQ * struct * * @param p_hwfn * * @return struct qed_eq* - a newly allocated structure; NULL upon error. */ struct qed_consq *qed_consq_alloc(struct qed_hwfn *p_hwfn); /** * @brief qed_consq_setup - Reset the ConsQ to its start * state. * * @param p_hwfn * @param p_eq */ void qed_consq_setup(struct qed_hwfn *p_hwfn, struct qed_consq *p_consq); /** * @brief qed_consq_free - deallocates the given ConsQ struct. * * @param p_hwfn * @param p_eq */ void qed_consq_free(struct qed_hwfn *p_hwfn, struct qed_consq *p_consq); /** * @file * * @brief Slow-hwfn low-level commands (Ramrods) function definitions. */ #define QED_SP_EQ_COMPLETION 0x01 #define QED_SP_CQE_COMPLETION 0x02 struct qed_sp_init_data { u32 cid; u16 opaque_fid; /* Information regarding operation upon sending & completion */ enum spq_mode comp_mode; struct qed_spq_comp_cb *p_comp_data; }; int qed_sp_init_request(struct qed_hwfn *p_hwfn, struct qed_spq_entry **pp_ent, u8 cmd, u8 protocol, struct qed_sp_init_data *p_data); /** * @brief qed_sp_pf_start - PF Function Start Ramrod * * This ramrod is sent to initialize a physical function (PF). It will * configure the function related parameters and write its completion to the * event ring specified in the parameters. * * Ramrods complete on the common event ring for the PF. This ring is * allocated by the driver on host memory and its parameters are written * to the internal RAM of the UStorm by the Function Start Ramrod. * * @param p_hwfn * @param p_tunn * @param mode * @param allow_npar_tx_switch * * @return int */ int qed_sp_pf_start(struct qed_hwfn *p_hwfn, struct qed_tunn_start_params *p_tunn, enum qed_mf_mode mode, bool allow_npar_tx_switch); /** * @brief qed_sp_pf_update - PF Function Update Ramrod * * This ramrod updates function-related parameters. Every parameter can be * updated independently, according to configuration flags. * * @param p_hwfn * * @return int */ int qed_sp_pf_update(struct qed_hwfn *p_hwfn); /** * @brief qed_sp_pf_stop - PF Function Stop Ramrod * * This ramrod is sent to close a Physical Function (PF). It is the last ramrod * sent and the last completion written to the PFs Event Ring. This ramrod also * deletes the context for the Slowhwfn connection on this PF. * * @note Not required for first packet. * * @param p_hwfn * * @return int */ int qed_sp_pf_stop(struct qed_hwfn *p_hwfn); int qed_sp_pf_update_tunn_cfg(struct qed_hwfn *p_hwfn, struct qed_tunn_update_params *p_tunn, enum spq_mode comp_mode, struct qed_spq_comp_cb *p_comp_data); /** * @brief qed_sp_heartbeat_ramrod - Send empty Ramrod * * @param p_hwfn * * @return int */ int qed_sp_heartbeat_ramrod(struct qed_hwfn *p_hwfn); #endif