| /* |
| * This file is part of the Chelsio FCoE driver for Linux. |
| * |
| * Copyright (c) 2008-2012 Chelsio Communications, Inc. All rights reserved. |
| * |
| * This software is available to you under a choice of one of two |
| * licenses. You may choose to be licensed under the terms of the GNU |
| * General Public License (GPL) Version 2, available from the file |
| * COPYING in the main directory of this source tree, or the |
| * OpenIB.org BSD license below: |
| * |
| * Redistribution and use in source and binary forms, with or |
| * without modification, are permitted provided that the following |
| * conditions are met: |
| * |
| * - Redistributions of source code must retain the above |
| * copyright notice, this list of conditions and the following |
| * disclaimer. |
| * |
| * - Redistributions in binary form must reproduce the above |
| * copyright notice, this list of conditions and the following |
| * disclaimer in the documentation and/or other materials |
| * provided with the distribution. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
| * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
| * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS |
| * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN |
| * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
| * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
| * SOFTWARE. |
| */ |
| |
| #ifndef __CSIO_WR_H__ |
| #define __CSIO_WR_H__ |
| |
| #include <linux/cache.h> |
| |
| #include "csio_defs.h" |
| #include "t4fw_api.h" |
| #include "t4fw_api_stor.h" |
| |
| /* |
| * SGE register field values. |
| */ |
| #define X_INGPCIEBOUNDARY_32B 0 |
| #define X_INGPCIEBOUNDARY_64B 1 |
| #define X_INGPCIEBOUNDARY_128B 2 |
| #define X_INGPCIEBOUNDARY_256B 3 |
| #define X_INGPCIEBOUNDARY_512B 4 |
| #define X_INGPCIEBOUNDARY_1024B 5 |
| #define X_INGPCIEBOUNDARY_2048B 6 |
| #define X_INGPCIEBOUNDARY_4096B 7 |
| |
| /* GTS register */ |
| #define X_TIMERREG_COUNTER0 0 |
| #define X_TIMERREG_COUNTER1 1 |
| #define X_TIMERREG_COUNTER2 2 |
| #define X_TIMERREG_COUNTER3 3 |
| #define X_TIMERREG_COUNTER4 4 |
| #define X_TIMERREG_COUNTER5 5 |
| #define X_TIMERREG_RESTART_COUNTER 6 |
| #define X_TIMERREG_UPDATE_CIDX 7 |
| |
| /* |
| * Egress Context field values |
| */ |
| #define X_FETCHBURSTMIN_16B 0 |
| #define X_FETCHBURSTMIN_32B 1 |
| #define X_FETCHBURSTMIN_64B 2 |
| #define X_FETCHBURSTMIN_128B 3 |
| |
| #define X_FETCHBURSTMAX_64B 0 |
| #define X_FETCHBURSTMAX_128B 1 |
| #define X_FETCHBURSTMAX_256B 2 |
| #define X_FETCHBURSTMAX_512B 3 |
| |
| #define X_HOSTFCMODE_NONE 0 |
| #define X_HOSTFCMODE_INGRESS_QUEUE 1 |
| #define X_HOSTFCMODE_STATUS_PAGE 2 |
| #define X_HOSTFCMODE_BOTH 3 |
| |
| /* |
| * Ingress Context field values |
| */ |
| #define X_UPDATESCHEDULING_TIMER 0 |
| #define X_UPDATESCHEDULING_COUNTER_OPTTIMER 1 |
| |
| #define X_UPDATEDELIVERY_NONE 0 |
| #define X_UPDATEDELIVERY_INTERRUPT 1 |
| #define X_UPDATEDELIVERY_STATUS_PAGE 2 |
| #define X_UPDATEDELIVERY_BOTH 3 |
| |
| #define X_INTERRUPTDESTINATION_PCIE 0 |
| #define X_INTERRUPTDESTINATION_IQ 1 |
| |
| #define X_RSPD_TYPE_FLBUF 0 |
| #define X_RSPD_TYPE_CPL 1 |
| #define X_RSPD_TYPE_INTR 2 |
| |
| /* WR status is at the same position as retval in a CMD header */ |
| #define csio_wr_status(_wr) \ |
| (FW_CMD_RETVAL_GET(ntohl(((struct fw_cmd_hdr *)(_wr))->lo))) |
| |
| struct csio_hw; |
| |
| extern int csio_intr_coalesce_cnt; |
| extern int csio_intr_coalesce_time; |
| |
| /* Ingress queue params */ |
| struct csio_iq_params { |
| |
| uint8_t iq_start:1; |
| uint8_t iq_stop:1; |
| uint8_t pfn:3; |
| |
| uint8_t vfn; |
| |
| uint16_t physiqid; |
| uint16_t iqid; |
| |
| uint16_t fl0id; |
| uint16_t fl1id; |
| |
| uint8_t viid; |
| |
| uint8_t type; |
| uint8_t iqasynch; |
| uint8_t reserved4; |
| |
| uint8_t iqandst; |
| uint8_t iqanus; |
| uint8_t iqanud; |
| |
| uint16_t iqandstindex; |
| |
| uint8_t iqdroprss; |
| uint8_t iqpciech; |
| uint8_t iqdcaen; |
| |
| uint8_t iqdcacpu; |
| uint8_t iqintcntthresh; |
| uint8_t iqo; |
| |
| uint8_t iqcprio; |
| uint8_t iqesize; |
| |
| uint16_t iqsize; |
| |
| uint64_t iqaddr; |
| |
| uint8_t iqflintiqhsen; |
| uint8_t reserved5; |
| uint8_t iqflintcongen; |
| uint8_t iqflintcngchmap; |
| |
| uint32_t reserved6; |
| |
| uint8_t fl0hostfcmode; |
| uint8_t fl0cprio; |
| uint8_t fl0paden; |
| uint8_t fl0packen; |
| uint8_t fl0congen; |
| uint8_t fl0dcaen; |
| |
| uint8_t fl0dcacpu; |
| uint8_t fl0fbmin; |
| |
| uint8_t fl0fbmax; |
| uint8_t fl0cidxfthresho; |
| uint8_t fl0cidxfthresh; |
| |
| uint16_t fl0size; |
| |
| uint64_t fl0addr; |
| |
| uint64_t reserved7; |
| |
| uint8_t fl1hostfcmode; |
| uint8_t fl1cprio; |
| uint8_t fl1paden; |
| uint8_t fl1packen; |
| uint8_t fl1congen; |
| uint8_t fl1dcaen; |
| |
| uint8_t fl1dcacpu; |
| uint8_t fl1fbmin; |
| |
| uint8_t fl1fbmax; |
| uint8_t fl1cidxfthresho; |
| uint8_t fl1cidxfthresh; |
| |
| uint16_t fl1size; |
| |
| uint64_t fl1addr; |
| }; |
| |
| /* Egress queue params */ |
| struct csio_eq_params { |
| |
| uint8_t pfn; |
| uint8_t vfn; |
| |
| uint8_t eqstart:1; |
| uint8_t eqstop:1; |
| |
| uint16_t physeqid; |
| uint32_t eqid; |
| |
| uint8_t hostfcmode:2; |
| uint8_t cprio:1; |
| uint8_t pciechn:3; |
| |
| uint16_t iqid; |
| |
| uint8_t dcaen:1; |
| uint8_t dcacpu:5; |
| |
| uint8_t fbmin:3; |
| uint8_t fbmax:3; |
| |
| uint8_t cidxfthresho:1; |
| uint8_t cidxfthresh:3; |
| |
| uint16_t eqsize; |
| |
| uint64_t eqaddr; |
| }; |
| |
| struct csio_dma_buf { |
| struct list_head list; |
| void *vaddr; /* Virtual address */ |
| dma_addr_t paddr; /* Physical address */ |
| uint32_t len; /* Buffer size */ |
| }; |
| |
| /* Generic I/O request structure */ |
| struct csio_ioreq { |
| struct csio_sm sm; /* SM, List |
| * should be the first member |
| */ |
| int iq_idx; /* Ingress queue index */ |
| int eq_idx; /* Egress queue index */ |
| uint32_t nsge; /* Number of SG elements */ |
| uint32_t tmo; /* Driver timeout */ |
| uint32_t datadir; /* Data direction */ |
| struct csio_dma_buf dma_buf; /* Req/resp DMA buffers */ |
| uint16_t wr_status; /* WR completion status */ |
| int16_t drv_status; /* Driver internal status */ |
| struct csio_lnode *lnode; /* Owner lnode */ |
| struct csio_rnode *rnode; /* Src/destination rnode */ |
| void (*io_cbfn) (struct csio_hw *, struct csio_ioreq *); |
| /* completion callback */ |
| void *scratch1; /* Scratch area 1. |
| */ |
| void *scratch2; /* Scratch area 2. */ |
| struct list_head gen_list; /* Any list associated with |
| * this ioreq. |
| */ |
| uint64_t fw_handle; /* Unique handle passed |
| * to FW |
| */ |
| uint8_t dcopy; /* Data copy required */ |
| uint8_t reserved1; |
| uint16_t reserved2; |
| struct completion cmplobj; /* ioreq completion object */ |
| } ____cacheline_aligned_in_smp; |
| |
| /* |
| * Egress status page for egress cidx updates |
| */ |
| struct csio_qstatus_page { |
| __be32 qid; |
| __be16 cidx; |
| __be16 pidx; |
| }; |
| |
| |
| enum { |
| CSIO_MAX_FLBUF_PER_IQWR = 4, |
| CSIO_QCREDIT_SZ = 64, /* pidx/cidx increments |
| * in bytes |
| */ |
| CSIO_MAX_QID = 0xFFFF, |
| CSIO_MAX_IQ = 128, |
| |
| CSIO_SGE_NTIMERS = 6, |
| CSIO_SGE_NCOUNTERS = 4, |
| CSIO_SGE_FL_SIZE_REGS = 16, |
| }; |
| |
| /* Defines for type */ |
| enum { |
| CSIO_EGRESS = 1, |
| CSIO_INGRESS = 2, |
| CSIO_FREELIST = 3, |
| }; |
| |
| /* |
| * Structure for footer (last 2 flits) of Ingress Queue Entry. |
| */ |
| struct csio_iqwr_footer { |
| __be32 hdrbuflen_pidx; |
| __be32 pldbuflen_qid; |
| union { |
| u8 type_gen; |
| __be64 last_flit; |
| } u; |
| }; |
| |
| #define IQWRF_NEWBUF (1 << 31) |
| #define IQWRF_LEN_GET(x) (((x) >> 0) & 0x7fffffffU) |
| #define IQWRF_GEN_SHIFT 7 |
| #define IQWRF_TYPE_GET(x) (((x) >> 4) & 0x3U) |
| |
| |
| /* |
| * WR pair: |
| * ======== |
| * A WR can start towards the end of a queue, and then continue at the |
| * beginning, since the queue is considered to be circular. This will |
| * require a pair of address/len to be passed back to the caller - |
| * hence the Work request pair structure. |
| */ |
| struct csio_wr_pair { |
| void *addr1; |
| uint32_t size1; |
| void *addr2; |
| uint32_t size2; |
| }; |
| |
| /* |
| * The following structure is used by ingress processing to return the |
| * free list buffers to consumers. |
| */ |
| struct csio_fl_dma_buf { |
| struct csio_dma_buf flbufs[CSIO_MAX_FLBUF_PER_IQWR]; |
| /* Freelist DMA buffers */ |
| int offset; /* Offset within the |
| * first FL buf. |
| */ |
| uint32_t totlen; /* Total length */ |
| uint8_t defer_free; /* Free of buffer can |
| * deferred |
| */ |
| }; |
| |
| /* Data-types */ |
| typedef void (*iq_handler_t)(struct csio_hw *, void *, uint32_t, |
| struct csio_fl_dma_buf *, void *); |
| |
| struct csio_iq { |
| uint16_t iqid; /* Queue ID */ |
| uint16_t physiqid; /* Physical Queue ID */ |
| uint16_t genbit; /* Generation bit, |
| * initially set to 1 |
| */ |
| int flq_idx; /* Freelist queue index */ |
| iq_handler_t iq_intx_handler; /* IQ INTx handler routine */ |
| }; |
| |
| struct csio_eq { |
| uint16_t eqid; /* Qid */ |
| uint16_t physeqid; /* Physical Queue ID */ |
| uint8_t wrap[512]; /* Temp area for q-wrap around*/ |
| }; |
| |
| struct csio_fl { |
| uint16_t flid; /* Qid */ |
| uint16_t packen; /* Packing enabled? */ |
| int offset; /* Offset within FL buf */ |
| int sreg; /* Size register */ |
| struct csio_dma_buf *bufs; /* Free list buffer ptr array |
| * indexed using flq->cidx/pidx |
| */ |
| }; |
| |
| struct csio_qstats { |
| uint32_t n_tot_reqs; /* Total no. of Requests */ |
| uint32_t n_tot_rsps; /* Total no. of responses */ |
| uint32_t n_qwrap; /* Queue wraps */ |
| uint32_t n_eq_wr_split; /* Number of split EQ WRs */ |
| uint32_t n_qentry; /* Queue entry */ |
| uint32_t n_qempty; /* Queue empty */ |
| uint32_t n_qfull; /* Queue fulls */ |
| uint32_t n_rsp_unknown; /* Unknown response type */ |
| uint32_t n_stray_comp; /* Stray completion intr */ |
| uint32_t n_flq_refill; /* Number of FL refills */ |
| }; |
| |
| /* Queue metadata */ |
| struct csio_q { |
| uint16_t type; /* Type: Ingress/Egress/FL */ |
| uint16_t pidx; /* producer index */ |
| uint16_t cidx; /* consumer index */ |
| uint16_t inc_idx; /* Incremental index */ |
| uint32_t wr_sz; /* Size of all WRs in this q |
| * if fixed |
| */ |
| void *vstart; /* Base virtual address |
| * of queue |
| */ |
| void *vwrap; /* Virtual end address to |
| * wrap around at |
| */ |
| uint32_t credits; /* Size of queue in credits */ |
| void *owner; /* Owner */ |
| union { /* Queue contexts */ |
| struct csio_iq iq; |
| struct csio_eq eq; |
| struct csio_fl fl; |
| } un; |
| |
| dma_addr_t pstart; /* Base physical address of |
| * queue |
| */ |
| uint32_t portid; /* PCIE Channel */ |
| uint32_t size; /* Size of queue in bytes */ |
| struct csio_qstats stats; /* Statistics */ |
| } ____cacheline_aligned_in_smp; |
| |
| struct csio_sge { |
| uint32_t csio_fl_align; /* Calculated and cached |
| * for fast path |
| */ |
| uint32_t sge_control; /* padding, boundaries, |
| * lengths, etc. |
| */ |
| uint32_t sge_host_page_size; /* Host page size */ |
| uint32_t sge_fl_buf_size[CSIO_SGE_FL_SIZE_REGS]; |
| /* free list buffer sizes */ |
| uint16_t timer_val[CSIO_SGE_NTIMERS]; |
| uint8_t counter_val[CSIO_SGE_NCOUNTERS]; |
| }; |
| |
| /* Work request module */ |
| struct csio_wrm { |
| int num_q; /* Number of queues */ |
| struct csio_q **q_arr; /* Array of queue pointers |
| * allocated dynamically |
| * based on configured values |
| */ |
| uint32_t fw_iq_start; /* Start ID of IQ for this fn*/ |
| uint32_t fw_eq_start; /* Start ID of EQ for this fn*/ |
| struct csio_q *intr_map[CSIO_MAX_IQ]; |
| /* IQ-id to IQ map table. */ |
| int free_qidx; /* queue idx of free queue */ |
| struct csio_sge sge; /* SGE params */ |
| }; |
| |
| #define csio_get_q(__hw, __idx) ((__hw)->wrm.q_arr[__idx]) |
| #define csio_q_type(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->type) |
| #define csio_q_pidx(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->pidx) |
| #define csio_q_cidx(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->cidx) |
| #define csio_q_inc_idx(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->inc_idx) |
| #define csio_q_vstart(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->vstart) |
| #define csio_q_pstart(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->pstart) |
| #define csio_q_size(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->size) |
| #define csio_q_credits(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->credits) |
| #define csio_q_portid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->portid) |
| #define csio_q_wr_sz(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->wr_sz) |
| #define csio_q_iqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.iq.iqid) |
| #define csio_q_physiqid(__hw, __idx) \ |
| ((__hw)->wrm.q_arr[(__idx)]->un.iq.physiqid) |
| #define csio_q_iq_flq_idx(__hw, __idx) \ |
| ((__hw)->wrm.q_arr[(__idx)]->un.iq.flq_idx) |
| #define csio_q_eqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid) |
| #define csio_q_flid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.fl.flid) |
| |
| #define csio_q_physeqid(__hw, __idx) \ |
| ((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid) |
| #define csio_iq_has_fl(__iq) ((__iq)->un.iq.flq_idx != -1) |
| |
| #define csio_q_iq_to_flid(__hw, __iq_idx) \ |
| csio_q_flid((__hw), (__hw)->wrm.q_arr[(__iq_qidx)]->un.iq.flq_idx) |
| #define csio_q_set_intr_map(__hw, __iq_idx, __rel_iq_id) \ |
| (__hw)->wrm.intr_map[__rel_iq_id] = csio_get_q(__hw, __iq_idx) |
| #define csio_q_eq_wrap(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap) |
| |
| struct csio_mb; |
| |
| int csio_wr_alloc_q(struct csio_hw *, uint32_t, uint32_t, |
| uint16_t, void *, uint32_t, int, iq_handler_t); |
| int csio_wr_iq_create(struct csio_hw *, void *, int, |
| uint32_t, uint8_t, bool, |
| void (*)(struct csio_hw *, struct csio_mb *)); |
| int csio_wr_eq_create(struct csio_hw *, void *, int, int, uint8_t, |
| void (*)(struct csio_hw *, struct csio_mb *)); |
| int csio_wr_destroy_queues(struct csio_hw *, bool cmd); |
| |
| |
| int csio_wr_get(struct csio_hw *, int, uint32_t, |
| struct csio_wr_pair *); |
| void csio_wr_copy_to_wrp(void *, struct csio_wr_pair *, uint32_t, uint32_t); |
| int csio_wr_issue(struct csio_hw *, int, bool); |
| int csio_wr_process_iq(struct csio_hw *, struct csio_q *, |
| void (*)(struct csio_hw *, void *, |
| uint32_t, struct csio_fl_dma_buf *, |
| void *), |
| void *); |
| int csio_wr_process_iq_idx(struct csio_hw *, int, |
| void (*)(struct csio_hw *, void *, |
| uint32_t, struct csio_fl_dma_buf *, |
| void *), |
| void *); |
| |
| void csio_wr_sge_init(struct csio_hw *); |
| int csio_wrm_init(struct csio_wrm *, struct csio_hw *); |
| void csio_wrm_exit(struct csio_wrm *, struct csio_hw *); |
| |
| #endif /* ifndef __CSIO_WR_H__ */ |