| /* |
| * Core routines and tables shareable across OS platforms. |
| * |
| * Copyright (c) 1994-2002 Justin T. Gibbs. |
| * Copyright (c) 2000-2003 Adaptec Inc. |
| * All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions, and the following disclaimer, |
| * without modification. |
| * 2. Redistributions in binary form must reproduce at minimum a disclaimer |
| * substantially similar to the "NO WARRANTY" disclaimer below |
| * ("Disclaimer") and any redistribution must be conditioned upon |
| * including a substantially similar Disclaimer requirement for further |
| * binary redistribution. |
| * 3. Neither the names of the above-listed copyright holders nor the names |
| * of any contributors may be used to endorse or promote products derived |
| * from this software without specific prior written permission. |
| * |
| * Alternatively, this software may be distributed under the terms of the |
| * GNU General Public License ("GPL") version 2 as published by the Free |
| * Software Foundation. |
| * |
| * NO WARRANTY |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR |
| * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| * HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, OR CONSEQUENTIAL |
| * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS |
| * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) |
| * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, |
| * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING |
| * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| * POSSIBILITY OF SUCH DAMAGES. |
| * |
| * $Id: //depot/aic7xxx/aic7xxx/aic79xx.c#250 $ |
| */ |
| |
| #ifdef __linux__ |
| #include "aic79xx_osm.h" |
| #include "aic79xx_inline.h" |
| #include "aicasm/aicasm_insformat.h" |
| #else |
| #include <dev/aic7xxx/aic79xx_osm.h> |
| #include <dev/aic7xxx/aic79xx_inline.h> |
| #include <dev/aic7xxx/aicasm/aicasm_insformat.h> |
| #endif |
| |
| |
| /***************************** Lookup Tables **********************************/ |
| static const char *const ahd_chip_names[] = |
| { |
| "NONE", |
| "aic7901", |
| "aic7902", |
| "aic7901A" |
| }; |
| static const u_int num_chip_names = ARRAY_SIZE(ahd_chip_names); |
| |
| /* |
| * Hardware error codes. |
| */ |
| struct ahd_hard_error_entry { |
| uint8_t errno; |
| const char *errmesg; |
| }; |
| |
| static const struct ahd_hard_error_entry ahd_hard_errors[] = { |
| { DSCTMOUT, "Discard Timer has timed out" }, |
| { ILLOPCODE, "Illegal Opcode in sequencer program" }, |
| { SQPARERR, "Sequencer Parity Error" }, |
| { DPARERR, "Data-path Parity Error" }, |
| { MPARERR, "Scratch or SCB Memory Parity Error" }, |
| { CIOPARERR, "CIOBUS Parity Error" }, |
| }; |
| static const u_int num_errors = ARRAY_SIZE(ahd_hard_errors); |
| |
| static const struct ahd_phase_table_entry ahd_phase_table[] = |
| { |
| { P_DATAOUT, MSG_NOOP, "in Data-out phase" }, |
| { P_DATAIN, MSG_INITIATOR_DET_ERR, "in Data-in phase" }, |
| { P_DATAOUT_DT, MSG_NOOP, "in DT Data-out phase" }, |
| { P_DATAIN_DT, MSG_INITIATOR_DET_ERR, "in DT Data-in phase" }, |
| { P_COMMAND, MSG_NOOP, "in Command phase" }, |
| { P_MESGOUT, MSG_NOOP, "in Message-out phase" }, |
| { P_STATUS, MSG_INITIATOR_DET_ERR, "in Status phase" }, |
| { P_MESGIN, MSG_PARITY_ERROR, "in Message-in phase" }, |
| { P_BUSFREE, MSG_NOOP, "while idle" }, |
| { 0, MSG_NOOP, "in unknown phase" } |
| }; |
| |
| /* |
| * In most cases we only wish to itterate over real phases, so |
| * exclude the last element from the count. |
| */ |
| static const u_int num_phases = ARRAY_SIZE(ahd_phase_table) - 1; |
| |
| /* Our Sequencer Program */ |
| #include "aic79xx_seq.h" |
| |
| /**************************** Function Declarations ***************************/ |
| static void ahd_handle_transmission_error(struct ahd_softc *ahd); |
| static void ahd_handle_lqiphase_error(struct ahd_softc *ahd, |
| u_int lqistat1); |
| static int ahd_handle_pkt_busfree(struct ahd_softc *ahd, |
| u_int busfreetime); |
| static int ahd_handle_nonpkt_busfree(struct ahd_softc *ahd); |
| static void ahd_handle_proto_violation(struct ahd_softc *ahd); |
| static void ahd_force_renegotiation(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo); |
| |
| static struct ahd_tmode_tstate* |
| ahd_alloc_tstate(struct ahd_softc *ahd, |
| u_int scsi_id, char channel); |
| #ifdef AHD_TARGET_MODE |
| static void ahd_free_tstate(struct ahd_softc *ahd, |
| u_int scsi_id, char channel, int force); |
| #endif |
| static void ahd_devlimited_syncrate(struct ahd_softc *ahd, |
| struct ahd_initiator_tinfo *, |
| u_int *period, |
| u_int *ppr_options, |
| role_t role); |
| static void ahd_update_neg_table(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo, |
| struct ahd_transinfo *tinfo); |
| static void ahd_update_pending_scbs(struct ahd_softc *ahd); |
| static void ahd_fetch_devinfo(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo); |
| static void ahd_scb_devinfo(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo, |
| struct scb *scb); |
| static void ahd_setup_initiator_msgout(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo, |
| struct scb *scb); |
| static void ahd_build_transfer_msg(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo); |
| static void ahd_construct_sdtr(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo, |
| u_int period, u_int offset); |
| static void ahd_construct_wdtr(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo, |
| u_int bus_width); |
| static void ahd_construct_ppr(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo, |
| u_int period, u_int offset, |
| u_int bus_width, u_int ppr_options); |
| static void ahd_clear_msg_state(struct ahd_softc *ahd); |
| static void ahd_handle_message_phase(struct ahd_softc *ahd); |
| typedef enum { |
| AHDMSG_1B, |
| AHDMSG_2B, |
| AHDMSG_EXT |
| } ahd_msgtype; |
| static int ahd_sent_msg(struct ahd_softc *ahd, ahd_msgtype type, |
| u_int msgval, int full); |
| static int ahd_parse_msg(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo); |
| static int ahd_handle_msg_reject(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo); |
| static void ahd_handle_ign_wide_residue(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo); |
| static void ahd_reinitialize_dataptrs(struct ahd_softc *ahd); |
| static void ahd_handle_devreset(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo, |
| u_int lun, cam_status status, |
| char *message, int verbose_level); |
| #ifdef AHD_TARGET_MODE |
| static void ahd_setup_target_msgin(struct ahd_softc *ahd, |
| struct ahd_devinfo *devinfo, |
| struct scb *scb); |
| #endif |
| |
| static u_int ahd_sglist_size(struct ahd_softc *ahd); |
| static u_int ahd_sglist_allocsize(struct ahd_softc *ahd); |
| static bus_dmamap_callback_t |
| ahd_dmamap_cb; |
| static void ahd_initialize_hscbs(struct ahd_softc *ahd); |
| static int ahd_init_scbdata(struct ahd_softc *ahd); |
| static void ahd_fini_scbdata(struct ahd_softc *ahd); |
| static void ahd_setup_iocell_workaround(struct ahd_softc *ahd); |
| static void ahd_iocell_first_selection(struct ahd_softc *ahd); |
| static void ahd_add_col_list(struct ahd_softc *ahd, |
| struct scb *scb, u_int col_idx); |
| static void ahd_rem_col_list(struct ahd_softc *ahd, |
| struct scb *scb); |
| static void ahd_chip_init(struct ahd_softc *ahd); |
| static void ahd_qinfifo_requeue(struct ahd_softc *ahd, |
| struct scb *prev_scb, |
| struct scb *scb); |
| static int ahd_qinfifo_count(struct ahd_softc *ahd); |
| static int ahd_search_scb_list(struct ahd_softc *ahd, int target, |
| char channel, int lun, u_int tag, |
| role_t role, uint32_t status, |
| ahd_search_action action, |
| u_int *list_head, u_int *list_tail, |
| u_int tid); |
| static void ahd_stitch_tid_list(struct ahd_softc *ahd, |
| u_int tid_prev, u_int tid_cur, |
| u_int tid_next); |
| static void ahd_add_scb_to_free_list(struct ahd_softc *ahd, |
| u_int scbid); |
| static u_int ahd_rem_wscb(struct ahd_softc *ahd, u_int scbid, |
| u_int prev, u_int next, u_int tid); |
| static void ahd_reset_current_bus(struct ahd_softc *ahd); |
| static ahd_callback_t ahd_stat_timer; |
| #ifdef AHD_DUMP_SEQ |
| static void ahd_dumpseq(struct ahd_softc *ahd); |
| #endif |
| static void ahd_loadseq(struct ahd_softc *ahd); |
| static int ahd_check_patch(struct ahd_softc *ahd, |
| const struct patch **start_patch, |
| u_int start_instr, u_int *skip_addr); |
| static u_int ahd_resolve_seqaddr(struct ahd_softc *ahd, |
| u_int address); |
| static void ahd_download_instr(struct ahd_softc *ahd, |
| u_int instrptr, uint8_t *dconsts); |
| static int ahd_probe_stack_size(struct ahd_softc *ahd); |
| static int ahd_scb_active_in_fifo(struct ahd_softc *ahd, |
| struct scb *scb); |
| static void ahd_run_data_fifo(struct ahd_softc *ahd, |
| struct scb *scb); |
| |
| #ifdef AHD_TARGET_MODE |
| static void ahd_queue_lstate_event(struct ahd_softc *ahd, |
| struct ahd_tmode_lstate *lstate, |
| u_int initiator_id, |
| u_int event_type, |
| u_int event_arg); |
| static void ahd_update_scsiid(struct ahd_softc *ahd, |
| u_int targid_mask); |
| static int ahd_handle_target_cmd(struct ahd_softc *ahd, |
| struct target_cmd *cmd); |
| #endif |
| |
| static int ahd_abort_scbs(struct ahd_softc *ahd, int target, |
| char channel, int lun, u_int tag, |
| role_t role, uint32_t status); |
| static void ahd_alloc_scbs(struct ahd_softc *ahd); |
| static void ahd_busy_tcl(struct ahd_softc *ahd, u_int tcl, |
| u_int scbid); |
| static void ahd_calc_residual(struct ahd_softc *ahd, |
| struct scb *scb); |
| static void ahd_clear_critical_section(struct ahd_softc *ahd); |
| static void ahd_clear_intstat(struct ahd_softc *ahd); |
| static void ahd_enable_coalescing(struct ahd_softc *ahd, |
| int enable); |
| static u_int ahd_find_busy_tcl(struct ahd_softc *ahd, u_int tcl); |
| static void ahd_freeze_devq(struct ahd_softc *ahd, |
| struct scb *scb); |
| static void ahd_handle_scb_status(struct ahd_softc *ahd, |
| struct scb *scb); |
| static const struct ahd_phase_table_entry* ahd_lookup_phase_entry(int phase); |
| static void ahd_shutdown(void *arg); |
| static void ahd_update_coalescing_values(struct ahd_softc *ahd, |
| u_int timer, |
| u_int maxcmds, |
| u_int mincmds); |
| static int ahd_verify_vpd_cksum(struct vpd_config *vpd); |
| static int ahd_wait_seeprom(struct ahd_softc *ahd); |
| static int ahd_match_scb(struct ahd_softc *ahd, struct scb *scb, |
| int target, char channel, int lun, |
| u_int tag, role_t role); |
| |
| static void ahd_reset_cmds_pending(struct ahd_softc *ahd); |
| |
| /*************************** Interrupt Services *******************************/ |
| static void ahd_run_qoutfifo(struct ahd_softc *ahd); |
| #ifdef AHD_TARGET_MODE |
| static void ahd_run_tqinfifo(struct ahd_softc *ahd, int paused); |
| #endif |
| static void ahd_handle_hwerrint(struct ahd_softc *ahd); |
| static void ahd_handle_seqint(struct ahd_softc *ahd, u_int intstat); |
| static void ahd_handle_scsiint(struct ahd_softc *ahd, |
| u_int intstat); |
| |
| /************************ Sequencer Execution Control *************************/ |
| void |
| ahd_set_modes(struct ahd_softc *ahd, ahd_mode src, ahd_mode dst) |
| { |
| if (ahd->src_mode == src && ahd->dst_mode == dst) |
| return; |
| #ifdef AHD_DEBUG |
| if (ahd->src_mode == AHD_MODE_UNKNOWN |
| || ahd->dst_mode == AHD_MODE_UNKNOWN) |
| panic("Setting mode prior to saving it.\n"); |
| if ((ahd_debug & AHD_SHOW_MODEPTR) != 0) |
| printk("%s: Setting mode 0x%x\n", ahd_name(ahd), |
| ahd_build_mode_state(ahd, src, dst)); |
| #endif |
| ahd_outb(ahd, MODE_PTR, ahd_build_mode_state(ahd, src, dst)); |
| ahd->src_mode = src; |
| ahd->dst_mode = dst; |
| } |
| |
| static void |
| ahd_update_modes(struct ahd_softc *ahd) |
| { |
| ahd_mode_state mode_ptr; |
| ahd_mode src; |
| ahd_mode dst; |
| |
| mode_ptr = ahd_inb(ahd, MODE_PTR); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MODEPTR) != 0) |
| printk("Reading mode 0x%x\n", mode_ptr); |
| #endif |
| ahd_extract_mode_state(ahd, mode_ptr, &src, &dst); |
| ahd_known_modes(ahd, src, dst); |
| } |
| |
| static void |
| ahd_assert_modes(struct ahd_softc *ahd, ahd_mode srcmode, |
| ahd_mode dstmode, const char *file, int line) |
| { |
| #ifdef AHD_DEBUG |
| if ((srcmode & AHD_MK_MSK(ahd->src_mode)) == 0 |
| || (dstmode & AHD_MK_MSK(ahd->dst_mode)) == 0) { |
| panic("%s:%s:%d: Mode assertion failed.\n", |
| ahd_name(ahd), file, line); |
| } |
| #endif |
| } |
| |
| #define AHD_ASSERT_MODES(ahd, source, dest) \ |
| ahd_assert_modes(ahd, source, dest, __FILE__, __LINE__); |
| |
| ahd_mode_state |
| ahd_save_modes(struct ahd_softc *ahd) |
| { |
| if (ahd->src_mode == AHD_MODE_UNKNOWN |
| || ahd->dst_mode == AHD_MODE_UNKNOWN) |
| ahd_update_modes(ahd); |
| |
| return (ahd_build_mode_state(ahd, ahd->src_mode, ahd->dst_mode)); |
| } |
| |
| void |
| ahd_restore_modes(struct ahd_softc *ahd, ahd_mode_state state) |
| { |
| ahd_mode src; |
| ahd_mode dst; |
| |
| ahd_extract_mode_state(ahd, state, &src, &dst); |
| ahd_set_modes(ahd, src, dst); |
| } |
| |
| /* |
| * Determine whether the sequencer has halted code execution. |
| * Returns non-zero status if the sequencer is stopped. |
| */ |
| int |
| ahd_is_paused(struct ahd_softc *ahd) |
| { |
| return ((ahd_inb(ahd, HCNTRL) & PAUSE) != 0); |
| } |
| |
| /* |
| * Request that the sequencer stop and wait, indefinitely, for it |
| * to stop. The sequencer will only acknowledge that it is paused |
| * once it has reached an instruction boundary and PAUSEDIS is |
| * cleared in the SEQCTL register. The sequencer may use PAUSEDIS |
| * for critical sections. |
| */ |
| void |
| ahd_pause(struct ahd_softc *ahd) |
| { |
| ahd_outb(ahd, HCNTRL, ahd->pause); |
| |
| /* |
| * Since the sequencer can disable pausing in a critical section, we |
| * must loop until it actually stops. |
| */ |
| while (ahd_is_paused(ahd) == 0) |
| ; |
| } |
| |
| /* |
| * Allow the sequencer to continue program execution. |
| * We check here to ensure that no additional interrupt |
| * sources that would cause the sequencer to halt have been |
| * asserted. If, for example, a SCSI bus reset is detected |
| * while we are fielding a different, pausing, interrupt type, |
| * we don't want to release the sequencer before going back |
| * into our interrupt handler and dealing with this new |
| * condition. |
| */ |
| void |
| ahd_unpause(struct ahd_softc *ahd) |
| { |
| /* |
| * Automatically restore our modes to those saved |
| * prior to the first change of the mode. |
| */ |
| if (ahd->saved_src_mode != AHD_MODE_UNKNOWN |
| && ahd->saved_dst_mode != AHD_MODE_UNKNOWN) { |
| if ((ahd->flags & AHD_UPDATE_PEND_CMDS) != 0) |
| ahd_reset_cmds_pending(ahd); |
| ahd_set_modes(ahd, ahd->saved_src_mode, ahd->saved_dst_mode); |
| } |
| |
| if ((ahd_inb(ahd, INTSTAT) & ~CMDCMPLT) == 0) |
| ahd_outb(ahd, HCNTRL, ahd->unpause); |
| |
| ahd_known_modes(ahd, AHD_MODE_UNKNOWN, AHD_MODE_UNKNOWN); |
| } |
| |
| /*********************** Scatter Gather List Handling *************************/ |
| void * |
| ahd_sg_setup(struct ahd_softc *ahd, struct scb *scb, |
| void *sgptr, dma_addr_t addr, bus_size_t len, int last) |
| { |
| scb->sg_count++; |
| if (sizeof(dma_addr_t) > 4 |
| && (ahd->flags & AHD_64BIT_ADDRESSING) != 0) { |
| struct ahd_dma64_seg *sg; |
| |
| sg = (struct ahd_dma64_seg *)sgptr; |
| sg->addr = ahd_htole64(addr); |
| sg->len = ahd_htole32(len | (last ? AHD_DMA_LAST_SEG : 0)); |
| return (sg + 1); |
| } else { |
| struct ahd_dma_seg *sg; |
| |
| sg = (struct ahd_dma_seg *)sgptr; |
| sg->addr = ahd_htole32(addr & 0xFFFFFFFF); |
| sg->len = ahd_htole32(len | ((addr >> 8) & 0x7F000000) |
| | (last ? AHD_DMA_LAST_SEG : 0)); |
| return (sg + 1); |
| } |
| } |
| |
| static void |
| ahd_setup_scb_common(struct ahd_softc *ahd, struct scb *scb) |
| { |
| /* XXX Handle target mode SCBs. */ |
| scb->crc_retry_count = 0; |
| if ((scb->flags & SCB_PACKETIZED) != 0) { |
| /* XXX what about ACA?? It is type 4, but TAG_TYPE == 0x3. */ |
| scb->hscb->task_attribute = scb->hscb->control & SCB_TAG_TYPE; |
| } else { |
| if (ahd_get_transfer_length(scb) & 0x01) |
| scb->hscb->task_attribute = SCB_XFERLEN_ODD; |
| else |
| scb->hscb->task_attribute = 0; |
| } |
| |
| if (scb->hscb->cdb_len <= MAX_CDB_LEN_WITH_SENSE_ADDR |
| || (scb->hscb->cdb_len & SCB_CDB_LEN_PTR) != 0) |
| scb->hscb->shared_data.idata.cdb_plus_saddr.sense_addr = |
| ahd_htole32(scb->sense_busaddr); |
| } |
| |
| static void |
| ahd_setup_data_scb(struct ahd_softc *ahd, struct scb *scb) |
| { |
| /* |
| * Copy the first SG into the "current" data ponter area. |
| */ |
| if ((ahd->flags & AHD_64BIT_ADDRESSING) != 0) { |
| struct ahd_dma64_seg *sg; |
| |
| sg = (struct ahd_dma64_seg *)scb->sg_list; |
| scb->hscb->dataptr = sg->addr; |
| scb->hscb->datacnt = sg->len; |
| } else { |
| struct ahd_dma_seg *sg; |
| uint32_t *dataptr_words; |
| |
| sg = (struct ahd_dma_seg *)scb->sg_list; |
| dataptr_words = (uint32_t*)&scb->hscb->dataptr; |
| dataptr_words[0] = sg->addr; |
| dataptr_words[1] = 0; |
| if ((ahd->flags & AHD_39BIT_ADDRESSING) != 0) { |
| uint64_t high_addr; |
| |
| high_addr = ahd_le32toh(sg->len) & 0x7F000000; |
| scb->hscb->dataptr |= ahd_htole64(high_addr << 8); |
| } |
| scb->hscb->datacnt = sg->len; |
| } |
| /* |
| * Note where to find the SG entries in bus space. |
| * We also set the full residual flag which the |
| * sequencer will clear as soon as a data transfer |
| * occurs. |
| */ |
| scb->hscb->sgptr = ahd_htole32(scb->sg_list_busaddr|SG_FULL_RESID); |
| } |
| |
| static void |
| ahd_setup_noxfer_scb(struct ahd_softc *ahd, struct scb *scb) |
| { |
| scb->hscb->sgptr = ahd_htole32(SG_LIST_NULL); |
| scb->hscb->dataptr = 0; |
| scb->hscb->datacnt = 0; |
| } |
| |
| /************************** Memory mapping routines ***************************/ |
| static void * |
| ahd_sg_bus_to_virt(struct ahd_softc *ahd, struct scb *scb, uint32_t sg_busaddr) |
| { |
| dma_addr_t sg_offset; |
| |
| /* sg_list_phys points to entry 1, not 0 */ |
| sg_offset = sg_busaddr - (scb->sg_list_busaddr - ahd_sg_size(ahd)); |
| return ((uint8_t *)scb->sg_list + sg_offset); |
| } |
| |
| static uint32_t |
| ahd_sg_virt_to_bus(struct ahd_softc *ahd, struct scb *scb, void *sg) |
| { |
| dma_addr_t sg_offset; |
| |
| /* sg_list_phys points to entry 1, not 0 */ |
| sg_offset = ((uint8_t *)sg - (uint8_t *)scb->sg_list) |
| - ahd_sg_size(ahd); |
| |
| return (scb->sg_list_busaddr + sg_offset); |
| } |
| |
| static void |
| ahd_sync_scb(struct ahd_softc *ahd, struct scb *scb, int op) |
| { |
| ahd_dmamap_sync(ahd, ahd->scb_data.hscb_dmat, |
| scb->hscb_map->dmamap, |
| /*offset*/(uint8_t*)scb->hscb - scb->hscb_map->vaddr, |
| /*len*/sizeof(*scb->hscb), op); |
| } |
| |
| void |
| ahd_sync_sglist(struct ahd_softc *ahd, struct scb *scb, int op) |
| { |
| if (scb->sg_count == 0) |
| return; |
| |
| ahd_dmamap_sync(ahd, ahd->scb_data.sg_dmat, |
| scb->sg_map->dmamap, |
| /*offset*/scb->sg_list_busaddr - ahd_sg_size(ahd), |
| /*len*/ahd_sg_size(ahd) * scb->sg_count, op); |
| } |
| |
| static void |
| ahd_sync_sense(struct ahd_softc *ahd, struct scb *scb, int op) |
| { |
| ahd_dmamap_sync(ahd, ahd->scb_data.sense_dmat, |
| scb->sense_map->dmamap, |
| /*offset*/scb->sense_busaddr, |
| /*len*/AHD_SENSE_BUFSIZE, op); |
| } |
| |
| #ifdef AHD_TARGET_MODE |
| static uint32_t |
| ahd_targetcmd_offset(struct ahd_softc *ahd, u_int index) |
| { |
| return (((uint8_t *)&ahd->targetcmds[index]) |
| - (uint8_t *)ahd->qoutfifo); |
| } |
| #endif |
| |
| /*********************** Miscellaneous Support Functions ***********************/ |
| /* |
| * Return pointers to the transfer negotiation information |
| * for the specified our_id/remote_id pair. |
| */ |
| struct ahd_initiator_tinfo * |
| ahd_fetch_transinfo(struct ahd_softc *ahd, char channel, u_int our_id, |
| u_int remote_id, struct ahd_tmode_tstate **tstate) |
| { |
| /* |
| * Transfer data structures are stored from the perspective |
| * of the target role. Since the parameters for a connection |
| * in the initiator role to a given target are the same as |
| * when the roles are reversed, we pretend we are the target. |
| */ |
| if (channel == 'B') |
| our_id += 8; |
| *tstate = ahd->enabled_targets[our_id]; |
| return (&(*tstate)->transinfo[remote_id]); |
| } |
| |
| uint16_t |
| ahd_inw(struct ahd_softc *ahd, u_int port) |
| { |
| /* |
| * Read high byte first as some registers increment |
| * or have other side effects when the low byte is |
| * read. |
| */ |
| uint16_t r = ahd_inb(ahd, port+1) << 8; |
| return r | ahd_inb(ahd, port); |
| } |
| |
| void |
| ahd_outw(struct ahd_softc *ahd, u_int port, u_int value) |
| { |
| /* |
| * Write low byte first to accommodate registers |
| * such as PRGMCNT where the order maters. |
| */ |
| ahd_outb(ahd, port, value & 0xFF); |
| ahd_outb(ahd, port+1, (value >> 8) & 0xFF); |
| } |
| |
| uint32_t |
| ahd_inl(struct ahd_softc *ahd, u_int port) |
| { |
| return ((ahd_inb(ahd, port)) |
| | (ahd_inb(ahd, port+1) << 8) |
| | (ahd_inb(ahd, port+2) << 16) |
| | (ahd_inb(ahd, port+3) << 24)); |
| } |
| |
| void |
| ahd_outl(struct ahd_softc *ahd, u_int port, uint32_t value) |
| { |
| ahd_outb(ahd, port, (value) & 0xFF); |
| ahd_outb(ahd, port+1, ((value) >> 8) & 0xFF); |
| ahd_outb(ahd, port+2, ((value) >> 16) & 0xFF); |
| ahd_outb(ahd, port+3, ((value) >> 24) & 0xFF); |
| } |
| |
| uint64_t |
| ahd_inq(struct ahd_softc *ahd, u_int port) |
| { |
| return ((ahd_inb(ahd, port)) |
| | (ahd_inb(ahd, port+1) << 8) |
| | (ahd_inb(ahd, port+2) << 16) |
| | (ahd_inb(ahd, port+3) << 24) |
| | (((uint64_t)ahd_inb(ahd, port+4)) << 32) |
| | (((uint64_t)ahd_inb(ahd, port+5)) << 40) |
| | (((uint64_t)ahd_inb(ahd, port+6)) << 48) |
| | (((uint64_t)ahd_inb(ahd, port+7)) << 56)); |
| } |
| |
| void |
| ahd_outq(struct ahd_softc *ahd, u_int port, uint64_t value) |
| { |
| ahd_outb(ahd, port, value & 0xFF); |
| ahd_outb(ahd, port+1, (value >> 8) & 0xFF); |
| ahd_outb(ahd, port+2, (value >> 16) & 0xFF); |
| ahd_outb(ahd, port+3, (value >> 24) & 0xFF); |
| ahd_outb(ahd, port+4, (value >> 32) & 0xFF); |
| ahd_outb(ahd, port+5, (value >> 40) & 0xFF); |
| ahd_outb(ahd, port+6, (value >> 48) & 0xFF); |
| ahd_outb(ahd, port+7, (value >> 56) & 0xFF); |
| } |
| |
| u_int |
| ahd_get_scbptr(struct ahd_softc *ahd) |
| { |
| AHD_ASSERT_MODES(ahd, ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK), |
| ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK)); |
| return (ahd_inb(ahd, SCBPTR) | (ahd_inb(ahd, SCBPTR + 1) << 8)); |
| } |
| |
| void |
| ahd_set_scbptr(struct ahd_softc *ahd, u_int scbptr) |
| { |
| AHD_ASSERT_MODES(ahd, ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK), |
| ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK)); |
| ahd_outb(ahd, SCBPTR, scbptr & 0xFF); |
| ahd_outb(ahd, SCBPTR+1, (scbptr >> 8) & 0xFF); |
| } |
| |
| #if 0 /* unused */ |
| static u_int |
| ahd_get_hnscb_qoff(struct ahd_softc *ahd) |
| { |
| return (ahd_inw_atomic(ahd, HNSCB_QOFF)); |
| } |
| #endif |
| |
| static void |
| ahd_set_hnscb_qoff(struct ahd_softc *ahd, u_int value) |
| { |
| ahd_outw_atomic(ahd, HNSCB_QOFF, value); |
| } |
| |
| #if 0 /* unused */ |
| static u_int |
| ahd_get_hescb_qoff(struct ahd_softc *ahd) |
| { |
| return (ahd_inb(ahd, HESCB_QOFF)); |
| } |
| #endif |
| |
| static void |
| ahd_set_hescb_qoff(struct ahd_softc *ahd, u_int value) |
| { |
| ahd_outb(ahd, HESCB_QOFF, value); |
| } |
| |
| static u_int |
| ahd_get_snscb_qoff(struct ahd_softc *ahd) |
| { |
| u_int oldvalue; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_CCHAN_MSK, AHD_MODE_CCHAN_MSK); |
| oldvalue = ahd_inw(ahd, SNSCB_QOFF); |
| ahd_outw(ahd, SNSCB_QOFF, oldvalue); |
| return (oldvalue); |
| } |
| |
| static void |
| ahd_set_snscb_qoff(struct ahd_softc *ahd, u_int value) |
| { |
| AHD_ASSERT_MODES(ahd, AHD_MODE_CCHAN_MSK, AHD_MODE_CCHAN_MSK); |
| ahd_outw(ahd, SNSCB_QOFF, value); |
| } |
| |
| #if 0 /* unused */ |
| static u_int |
| ahd_get_sescb_qoff(struct ahd_softc *ahd) |
| { |
| AHD_ASSERT_MODES(ahd, AHD_MODE_CCHAN_MSK, AHD_MODE_CCHAN_MSK); |
| return (ahd_inb(ahd, SESCB_QOFF)); |
| } |
| #endif |
| |
| static void |
| ahd_set_sescb_qoff(struct ahd_softc *ahd, u_int value) |
| { |
| AHD_ASSERT_MODES(ahd, AHD_MODE_CCHAN_MSK, AHD_MODE_CCHAN_MSK); |
| ahd_outb(ahd, SESCB_QOFF, value); |
| } |
| |
| #if 0 /* unused */ |
| static u_int |
| ahd_get_sdscb_qoff(struct ahd_softc *ahd) |
| { |
| AHD_ASSERT_MODES(ahd, AHD_MODE_CCHAN_MSK, AHD_MODE_CCHAN_MSK); |
| return (ahd_inb(ahd, SDSCB_QOFF) | (ahd_inb(ahd, SDSCB_QOFF + 1) << 8)); |
| } |
| #endif |
| |
| static void |
| ahd_set_sdscb_qoff(struct ahd_softc *ahd, u_int value) |
| { |
| AHD_ASSERT_MODES(ahd, AHD_MODE_CCHAN_MSK, AHD_MODE_CCHAN_MSK); |
| ahd_outb(ahd, SDSCB_QOFF, value & 0xFF); |
| ahd_outb(ahd, SDSCB_QOFF+1, (value >> 8) & 0xFF); |
| } |
| |
| u_int |
| ahd_inb_scbram(struct ahd_softc *ahd, u_int offset) |
| { |
| u_int value; |
| |
| /* |
| * Workaround PCI-X Rev A. hardware bug. |
| * After a host read of SCB memory, the chip |
| * may become confused into thinking prefetch |
| * was required. This starts the discard timer |
| * running and can cause an unexpected discard |
| * timer interrupt. The work around is to read |
| * a normal register prior to the exhaustion of |
| * the discard timer. The mode pointer register |
| * has no side effects and so serves well for |
| * this purpose. |
| * |
| * Razor #528 |
| */ |
| value = ahd_inb(ahd, offset); |
| if ((ahd->bugs & AHD_PCIX_SCBRAM_RD_BUG) != 0) |
| ahd_inb(ahd, MODE_PTR); |
| return (value); |
| } |
| |
| u_int |
| ahd_inw_scbram(struct ahd_softc *ahd, u_int offset) |
| { |
| return (ahd_inb_scbram(ahd, offset) |
| | (ahd_inb_scbram(ahd, offset+1) << 8)); |
| } |
| |
| static uint32_t |
| ahd_inl_scbram(struct ahd_softc *ahd, u_int offset) |
| { |
| return (ahd_inw_scbram(ahd, offset) |
| | (ahd_inw_scbram(ahd, offset+2) << 16)); |
| } |
| |
| static uint64_t |
| ahd_inq_scbram(struct ahd_softc *ahd, u_int offset) |
| { |
| return (ahd_inl_scbram(ahd, offset) |
| | ((uint64_t)ahd_inl_scbram(ahd, offset+4)) << 32); |
| } |
| |
| struct scb * |
| ahd_lookup_scb(struct ahd_softc *ahd, u_int tag) |
| { |
| struct scb* scb; |
| |
| if (tag >= AHD_SCB_MAX) |
| return (NULL); |
| scb = ahd->scb_data.scbindex[tag]; |
| if (scb != NULL) |
| ahd_sync_scb(ahd, scb, |
| BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE); |
| return (scb); |
| } |
| |
| static void |
| ahd_swap_with_next_hscb(struct ahd_softc *ahd, struct scb *scb) |
| { |
| struct hardware_scb *q_hscb; |
| struct map_node *q_hscb_map; |
| uint32_t saved_hscb_busaddr; |
| |
| /* |
| * Our queuing method is a bit tricky. The card |
| * knows in advance which HSCB (by address) to download, |
| * and we can't disappoint it. To achieve this, the next |
| * HSCB to download is saved off in ahd->next_queued_hscb. |
| * When we are called to queue "an arbitrary scb", |
| * we copy the contents of the incoming HSCB to the one |
| * the sequencer knows about, swap HSCB pointers and |
| * finally assign the SCB to the tag indexed location |
| * in the scb_array. This makes sure that we can still |
| * locate the correct SCB by SCB_TAG. |
| */ |
| q_hscb = ahd->next_queued_hscb; |
| q_hscb_map = ahd->next_queued_hscb_map; |
| saved_hscb_busaddr = q_hscb->hscb_busaddr; |
| memcpy(q_hscb, scb->hscb, sizeof(*scb->hscb)); |
| q_hscb->hscb_busaddr = saved_hscb_busaddr; |
| q_hscb->next_hscb_busaddr = scb->hscb->hscb_busaddr; |
| |
| /* Now swap HSCB pointers. */ |
| ahd->next_queued_hscb = scb->hscb; |
| ahd->next_queued_hscb_map = scb->hscb_map; |
| scb->hscb = q_hscb; |
| scb->hscb_map = q_hscb_map; |
| |
| /* Now define the mapping from tag to SCB in the scbindex */ |
| ahd->scb_data.scbindex[SCB_GET_TAG(scb)] = scb; |
| } |
| |
| /* |
| * Tell the sequencer about a new transaction to execute. |
| */ |
| void |
| ahd_queue_scb(struct ahd_softc *ahd, struct scb *scb) |
| { |
| ahd_swap_with_next_hscb(ahd, scb); |
| |
| if (SCBID_IS_NULL(SCB_GET_TAG(scb))) |
| panic("Attempt to queue invalid SCB tag %x\n", |
| SCB_GET_TAG(scb)); |
| |
| /* |
| * Keep a history of SCBs we've downloaded in the qinfifo. |
| */ |
| ahd->qinfifo[AHD_QIN_WRAP(ahd->qinfifonext)] = SCB_GET_TAG(scb); |
| ahd->qinfifonext++; |
| |
| if (scb->sg_count != 0) |
| ahd_setup_data_scb(ahd, scb); |
| else |
| ahd_setup_noxfer_scb(ahd, scb); |
| ahd_setup_scb_common(ahd, scb); |
| |
| /* |
| * Make sure our data is consistent from the |
| * perspective of the adapter. |
| */ |
| ahd_sync_scb(ahd, scb, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_QUEUE) != 0) { |
| uint64_t host_dataptr; |
| |
| host_dataptr = ahd_le64toh(scb->hscb->dataptr); |
| printk("%s: Queueing SCB %d:0x%x bus addr 0x%x - 0x%x%x/0x%x\n", |
| ahd_name(ahd), |
| SCB_GET_TAG(scb), scb->hscb->scsiid, |
| ahd_le32toh(scb->hscb->hscb_busaddr), |
| (u_int)((host_dataptr >> 32) & 0xFFFFFFFF), |
| (u_int)(host_dataptr & 0xFFFFFFFF), |
| ahd_le32toh(scb->hscb->datacnt)); |
| } |
| #endif |
| /* Tell the adapter about the newly queued SCB */ |
| ahd_set_hnscb_qoff(ahd, ahd->qinfifonext); |
| } |
| |
| /************************** Interrupt Processing ******************************/ |
| static void |
| ahd_sync_qoutfifo(struct ahd_softc *ahd, int op) |
| { |
| ahd_dmamap_sync(ahd, ahd->shared_data_dmat, ahd->shared_data_map.dmamap, |
| /*offset*/0, |
| /*len*/AHD_SCB_MAX * sizeof(struct ahd_completion), op); |
| } |
| |
| static void |
| ahd_sync_tqinfifo(struct ahd_softc *ahd, int op) |
| { |
| #ifdef AHD_TARGET_MODE |
| if ((ahd->flags & AHD_TARGETROLE) != 0) { |
| ahd_dmamap_sync(ahd, ahd->shared_data_dmat, |
| ahd->shared_data_map.dmamap, |
| ahd_targetcmd_offset(ahd, 0), |
| sizeof(struct target_cmd) * AHD_TMODE_CMDS, |
| op); |
| } |
| #endif |
| } |
| |
| /* |
| * See if the firmware has posted any completed commands |
| * into our in-core command complete fifos. |
| */ |
| #define AHD_RUN_QOUTFIFO 0x1 |
| #define AHD_RUN_TQINFIFO 0x2 |
| static u_int |
| ahd_check_cmdcmpltqueues(struct ahd_softc *ahd) |
| { |
| u_int retval; |
| |
| retval = 0; |
| ahd_dmamap_sync(ahd, ahd->shared_data_dmat, ahd->shared_data_map.dmamap, |
| /*offset*/ahd->qoutfifonext * sizeof(*ahd->qoutfifo), |
| /*len*/sizeof(*ahd->qoutfifo), BUS_DMASYNC_POSTREAD); |
| if (ahd->qoutfifo[ahd->qoutfifonext].valid_tag |
| == ahd->qoutfifonext_valid_tag) |
| retval |= AHD_RUN_QOUTFIFO; |
| #ifdef AHD_TARGET_MODE |
| if ((ahd->flags & AHD_TARGETROLE) != 0 |
| && (ahd->flags & AHD_TQINFIFO_BLOCKED) == 0) { |
| ahd_dmamap_sync(ahd, ahd->shared_data_dmat, |
| ahd->shared_data_map.dmamap, |
| ahd_targetcmd_offset(ahd, ahd->tqinfifofnext), |
| /*len*/sizeof(struct target_cmd), |
| BUS_DMASYNC_POSTREAD); |
| if (ahd->targetcmds[ahd->tqinfifonext].cmd_valid != 0) |
| retval |= AHD_RUN_TQINFIFO; |
| } |
| #endif |
| return (retval); |
| } |
| |
| /* |
| * Catch an interrupt from the adapter |
| */ |
| int |
| ahd_intr(struct ahd_softc *ahd) |
| { |
| u_int intstat; |
| |
| if ((ahd->pause & INTEN) == 0) { |
| /* |
| * Our interrupt is not enabled on the chip |
| * and may be disabled for re-entrancy reasons, |
| * so just return. This is likely just a shared |
| * interrupt. |
| */ |
| return (0); |
| } |
| |
| /* |
| * Instead of directly reading the interrupt status register, |
| * infer the cause of the interrupt by checking our in-core |
| * completion queues. This avoids a costly PCI bus read in |
| * most cases. |
| */ |
| if ((ahd->flags & AHD_ALL_INTERRUPTS) == 0 |
| && (ahd_check_cmdcmpltqueues(ahd) != 0)) |
| intstat = CMDCMPLT; |
| else |
| intstat = ahd_inb(ahd, INTSTAT); |
| |
| if ((intstat & INT_PEND) == 0) |
| return (0); |
| |
| if (intstat & CMDCMPLT) { |
| ahd_outb(ahd, CLRINT, CLRCMDINT); |
| |
| /* |
| * Ensure that the chip sees that we've cleared |
| * this interrupt before we walk the output fifo. |
| * Otherwise, we may, due to posted bus writes, |
| * clear the interrupt after we finish the scan, |
| * and after the sequencer has added new entries |
| * and asserted the interrupt again. |
| */ |
| if ((ahd->bugs & AHD_INTCOLLISION_BUG) != 0) { |
| if (ahd_is_paused(ahd)) { |
| /* |
| * Potentially lost SEQINT. |
| * If SEQINTCODE is non-zero, |
| * simulate the SEQINT. |
| */ |
| if (ahd_inb(ahd, SEQINTCODE) != NO_SEQINT) |
| intstat |= SEQINT; |
| } |
| } else { |
| ahd_flush_device_writes(ahd); |
| } |
| ahd_run_qoutfifo(ahd); |
| ahd->cmdcmplt_counts[ahd->cmdcmplt_bucket]++; |
| ahd->cmdcmplt_total++; |
| #ifdef AHD_TARGET_MODE |
| if ((ahd->flags & AHD_TARGETROLE) != 0) |
| ahd_run_tqinfifo(ahd, /*paused*/FALSE); |
| #endif |
| } |
| |
| /* |
| * Handle statuses that may invalidate our cached |
| * copy of INTSTAT separately. |
| */ |
| if (intstat == 0xFF && (ahd->features & AHD_REMOVABLE) != 0) { |
| /* Hot eject. Do nothing */ |
| } else if (intstat & HWERRINT) { |
| ahd_handle_hwerrint(ahd); |
| } else if ((intstat & (PCIINT|SPLTINT)) != 0) { |
| ahd->bus_intr(ahd); |
| } else { |
| |
| if ((intstat & SEQINT) != 0) |
| ahd_handle_seqint(ahd, intstat); |
| |
| if ((intstat & SCSIINT) != 0) |
| ahd_handle_scsiint(ahd, intstat); |
| } |
| return (1); |
| } |
| |
| /******************************** Private Inlines *****************************/ |
| static inline void |
| ahd_assert_atn(struct ahd_softc *ahd) |
| { |
| ahd_outb(ahd, SCSISIGO, ATNO); |
| } |
| |
| /* |
| * Determine if the current connection has a packetized |
| * agreement. This does not necessarily mean that we |
| * are currently in a packetized transfer. We could |
| * just as easily be sending or receiving a message. |
| */ |
| static int |
| ahd_currently_packetized(struct ahd_softc *ahd) |
| { |
| ahd_mode_state saved_modes; |
| int packetized; |
| |
| saved_modes = ahd_save_modes(ahd); |
| if ((ahd->bugs & AHD_PKTIZED_STATUS_BUG) != 0) { |
| /* |
| * The packetized bit refers to the last |
| * connection, not the current one. Check |
| * for non-zero LQISTATE instead. |
| */ |
| ahd_set_modes(ahd, AHD_MODE_CFG, AHD_MODE_CFG); |
| packetized = ahd_inb(ahd, LQISTATE) != 0; |
| } else { |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| packetized = ahd_inb(ahd, LQISTAT2) & PACKETIZED; |
| } |
| ahd_restore_modes(ahd, saved_modes); |
| return (packetized); |
| } |
| |
| static inline int |
| ahd_set_active_fifo(struct ahd_softc *ahd) |
| { |
| u_int active_fifo; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| active_fifo = ahd_inb(ahd, DFFSTAT) & CURRFIFO; |
| switch (active_fifo) { |
| case 0: |
| case 1: |
| ahd_set_modes(ahd, active_fifo, active_fifo); |
| return (1); |
| default: |
| return (0); |
| } |
| } |
| |
| static inline void |
| ahd_unbusy_tcl(struct ahd_softc *ahd, u_int tcl) |
| { |
| ahd_busy_tcl(ahd, tcl, SCB_LIST_NULL); |
| } |
| |
| /* |
| * Determine whether the sequencer reported a residual |
| * for this SCB/transaction. |
| */ |
| static inline void |
| ahd_update_residual(struct ahd_softc *ahd, struct scb *scb) |
| { |
| uint32_t sgptr; |
| |
| sgptr = ahd_le32toh(scb->hscb->sgptr); |
| if ((sgptr & SG_STATUS_VALID) != 0) |
| ahd_calc_residual(ahd, scb); |
| } |
| |
| static inline void |
| ahd_complete_scb(struct ahd_softc *ahd, struct scb *scb) |
| { |
| uint32_t sgptr; |
| |
| sgptr = ahd_le32toh(scb->hscb->sgptr); |
| if ((sgptr & SG_STATUS_VALID) != 0) |
| ahd_handle_scb_status(ahd, scb); |
| else |
| ahd_done(ahd, scb); |
| } |
| |
| |
| /************************* Sequencer Execution Control ************************/ |
| /* |
| * Restart the sequencer program from address zero |
| */ |
| static void |
| ahd_restart(struct ahd_softc *ahd) |
| { |
| |
| ahd_pause(ahd); |
| |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| |
| /* No more pending messages */ |
| ahd_clear_msg_state(ahd); |
| ahd_outb(ahd, SCSISIGO, 0); /* De-assert BSY */ |
| ahd_outb(ahd, MSG_OUT, MSG_NOOP); /* No message to send */ |
| ahd_outb(ahd, SXFRCTL1, ahd_inb(ahd, SXFRCTL1) & ~BITBUCKET); |
| ahd_outb(ahd, SEQINTCTL, 0); |
| ahd_outb(ahd, LASTPHASE, P_BUSFREE); |
| ahd_outb(ahd, SEQ_FLAGS, 0); |
| ahd_outb(ahd, SAVED_SCSIID, 0xFF); |
| ahd_outb(ahd, SAVED_LUN, 0xFF); |
| |
| /* |
| * Ensure that the sequencer's idea of TQINPOS |
| * matches our own. The sequencer increments TQINPOS |
| * only after it sees a DMA complete and a reset could |
| * occur before the increment leaving the kernel to believe |
| * the command arrived but the sequencer to not. |
| */ |
| ahd_outb(ahd, TQINPOS, ahd->tqinfifonext); |
| |
| /* Always allow reselection */ |
| ahd_outb(ahd, SCSISEQ1, |
| ahd_inb(ahd, SCSISEQ_TEMPLATE) & (ENSELI|ENRSELI|ENAUTOATNP)); |
| ahd_set_modes(ahd, AHD_MODE_CCHAN, AHD_MODE_CCHAN); |
| |
| /* |
| * Clear any pending sequencer interrupt. It is no |
| * longer relevant since we're resetting the Program |
| * Counter. |
| */ |
| ahd_outb(ahd, CLRINT, CLRSEQINT); |
| |
| ahd_outb(ahd, SEQCTL0, FASTMODE|SEQRESET); |
| ahd_unpause(ahd); |
| } |
| |
| static void |
| ahd_clear_fifo(struct ahd_softc *ahd, u_int fifo) |
| { |
| ahd_mode_state saved_modes; |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_FIFOS) != 0) |
| printk("%s: Clearing FIFO %d\n", ahd_name(ahd), fifo); |
| #endif |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, fifo, fifo); |
| ahd_outb(ahd, DFFSXFRCTL, RSTCHN|CLRSHCNT); |
| if ((ahd_inb(ahd, SG_STATE) & FETCH_INPROG) != 0) |
| ahd_outb(ahd, CCSGCTL, CCSGRESET); |
| ahd_outb(ahd, LONGJMP_ADDR + 1, INVALID_ADDR); |
| ahd_outb(ahd, SG_STATE, 0); |
| ahd_restore_modes(ahd, saved_modes); |
| } |
| |
| /************************* Input/Output Queues ********************************/ |
| /* |
| * Flush and completed commands that are sitting in the command |
| * complete queues down on the chip but have yet to be dma'ed back up. |
| */ |
| static void |
| ahd_flush_qoutfifo(struct ahd_softc *ahd) |
| { |
| struct scb *scb; |
| ahd_mode_state saved_modes; |
| u_int saved_scbptr; |
| u_int ccscbctl; |
| u_int scbid; |
| u_int next_scbid; |
| |
| saved_modes = ahd_save_modes(ahd); |
| |
| /* |
| * Flush the good status FIFO for completed packetized commands. |
| */ |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| saved_scbptr = ahd_get_scbptr(ahd); |
| while ((ahd_inb(ahd, LQISTAT2) & LQIGSAVAIL) != 0) { |
| u_int fifo_mode; |
| u_int i; |
| |
| scbid = ahd_inw(ahd, GSFIFO); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| printk("%s: Warning - GSFIFO SCB %d invalid\n", |
| ahd_name(ahd), scbid); |
| continue; |
| } |
| /* |
| * Determine if this transaction is still active in |
| * any FIFO. If it is, we must flush that FIFO to |
| * the host before completing the command. |
| */ |
| fifo_mode = 0; |
| rescan_fifos: |
| for (i = 0; i < 2; i++) { |
| /* Toggle to the other mode. */ |
| fifo_mode ^= 1; |
| ahd_set_modes(ahd, fifo_mode, fifo_mode); |
| |
| if (ahd_scb_active_in_fifo(ahd, scb) == 0) |
| continue; |
| |
| ahd_run_data_fifo(ahd, scb); |
| |
| /* |
| * Running this FIFO may cause a CFG4DATA for |
| * this same transaction to assert in the other |
| * FIFO or a new snapshot SAVEPTRS interrupt |
| * in this FIFO. Even running a FIFO may not |
| * clear the transaction if we are still waiting |
| * for data to drain to the host. We must loop |
| * until the transaction is not active in either |
| * FIFO just to be sure. Reset our loop counter |
| * so we will visit both FIFOs again before |
| * declaring this transaction finished. We |
| * also delay a bit so that status has a chance |
| * to change before we look at this FIFO again. |
| */ |
| ahd_delay(200); |
| goto rescan_fifos; |
| } |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| ahd_set_scbptr(ahd, scbid); |
| if ((ahd_inb_scbram(ahd, SCB_SGPTR) & SG_LIST_NULL) == 0 |
| && ((ahd_inb_scbram(ahd, SCB_SGPTR) & SG_FULL_RESID) != 0 |
| || (ahd_inb_scbram(ahd, SCB_RESIDUAL_SGPTR) |
| & SG_LIST_NULL) != 0)) { |
| u_int comp_head; |
| |
| /* |
| * The transfer completed with a residual. |
| * Place this SCB on the complete DMA list |
| * so that we update our in-core copy of the |
| * SCB before completing the command. |
| */ |
| ahd_outb(ahd, SCB_SCSI_STATUS, 0); |
| ahd_outb(ahd, SCB_SGPTR, |
| ahd_inb_scbram(ahd, SCB_SGPTR) |
| | SG_STATUS_VALID); |
| ahd_outw(ahd, SCB_TAG, scbid); |
| ahd_outw(ahd, SCB_NEXT_COMPLETE, SCB_LIST_NULL); |
| comp_head = ahd_inw(ahd, COMPLETE_DMA_SCB_HEAD); |
| if (SCBID_IS_NULL(comp_head)) { |
| ahd_outw(ahd, COMPLETE_DMA_SCB_HEAD, scbid); |
| ahd_outw(ahd, COMPLETE_DMA_SCB_TAIL, scbid); |
| } else { |
| u_int tail; |
| |
| tail = ahd_inw(ahd, COMPLETE_DMA_SCB_TAIL); |
| ahd_set_scbptr(ahd, tail); |
| ahd_outw(ahd, SCB_NEXT_COMPLETE, scbid); |
| ahd_outw(ahd, COMPLETE_DMA_SCB_TAIL, scbid); |
| ahd_set_scbptr(ahd, scbid); |
| } |
| } else |
| ahd_complete_scb(ahd, scb); |
| } |
| ahd_set_scbptr(ahd, saved_scbptr); |
| |
| /* |
| * Setup for command channel portion of flush. |
| */ |
| ahd_set_modes(ahd, AHD_MODE_CCHAN, AHD_MODE_CCHAN); |
| |
| /* |
| * Wait for any inprogress DMA to complete and clear DMA state |
| * if this is for an SCB in the qinfifo. |
| */ |
| while (((ccscbctl = ahd_inb(ahd, CCSCBCTL)) & (CCARREN|CCSCBEN)) != 0) { |
| |
| if ((ccscbctl & (CCSCBDIR|CCARREN)) == (CCSCBDIR|CCARREN)) { |
| if ((ccscbctl & ARRDONE) != 0) |
| break; |
| } else if ((ccscbctl & CCSCBDONE) != 0) |
| break; |
| ahd_delay(200); |
| } |
| /* |
| * We leave the sequencer to cleanup in the case of DMA's to |
| * update the qoutfifo. In all other cases (DMA's to the |
| * chip or a push of an SCB from the COMPLETE_DMA_SCB list), |
| * we disable the DMA engine so that the sequencer will not |
| * attempt to handle the DMA completion. |
| */ |
| if ((ccscbctl & CCSCBDIR) != 0 || (ccscbctl & ARRDONE) != 0) |
| ahd_outb(ahd, CCSCBCTL, ccscbctl & ~(CCARREN|CCSCBEN)); |
| |
| /* |
| * Complete any SCBs that just finished |
| * being DMA'ed into the qoutfifo. |
| */ |
| ahd_run_qoutfifo(ahd); |
| |
| saved_scbptr = ahd_get_scbptr(ahd); |
| /* |
| * Manually update/complete any completed SCBs that are waiting to be |
| * DMA'ed back up to the host. |
| */ |
| scbid = ahd_inw(ahd, COMPLETE_DMA_SCB_HEAD); |
| while (!SCBID_IS_NULL(scbid)) { |
| uint8_t *hscb_ptr; |
| u_int i; |
| |
| ahd_set_scbptr(ahd, scbid); |
| next_scbid = ahd_inw_scbram(ahd, SCB_NEXT_COMPLETE); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| printk("%s: Warning - DMA-up and complete " |
| "SCB %d invalid\n", ahd_name(ahd), scbid); |
| continue; |
| } |
| hscb_ptr = (uint8_t *)scb->hscb; |
| for (i = 0; i < sizeof(struct hardware_scb); i++) |
| *hscb_ptr++ = ahd_inb_scbram(ahd, SCB_BASE + i); |
| |
| ahd_complete_scb(ahd, scb); |
| scbid = next_scbid; |
| } |
| ahd_outw(ahd, COMPLETE_DMA_SCB_HEAD, SCB_LIST_NULL); |
| ahd_outw(ahd, COMPLETE_DMA_SCB_TAIL, SCB_LIST_NULL); |
| |
| scbid = ahd_inw(ahd, COMPLETE_ON_QFREEZE_HEAD); |
| while (!SCBID_IS_NULL(scbid)) { |
| |
| ahd_set_scbptr(ahd, scbid); |
| next_scbid = ahd_inw_scbram(ahd, SCB_NEXT_COMPLETE); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| printk("%s: Warning - Complete Qfrz SCB %d invalid\n", |
| ahd_name(ahd), scbid); |
| continue; |
| } |
| |
| ahd_complete_scb(ahd, scb); |
| scbid = next_scbid; |
| } |
| ahd_outw(ahd, COMPLETE_ON_QFREEZE_HEAD, SCB_LIST_NULL); |
| |
| scbid = ahd_inw(ahd, COMPLETE_SCB_HEAD); |
| while (!SCBID_IS_NULL(scbid)) { |
| |
| ahd_set_scbptr(ahd, scbid); |
| next_scbid = ahd_inw_scbram(ahd, SCB_NEXT_COMPLETE); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| printk("%s: Warning - Complete SCB %d invalid\n", |
| ahd_name(ahd), scbid); |
| continue; |
| } |
| |
| ahd_complete_scb(ahd, scb); |
| scbid = next_scbid; |
| } |
| ahd_outw(ahd, COMPLETE_SCB_HEAD, SCB_LIST_NULL); |
| |
| /* |
| * Restore state. |
| */ |
| ahd_set_scbptr(ahd, saved_scbptr); |
| ahd_restore_modes(ahd, saved_modes); |
| ahd->flags |= AHD_UPDATE_PEND_CMDS; |
| } |
| |
| /* |
| * Determine if an SCB for a packetized transaction |
| * is active in a FIFO. |
| */ |
| static int |
| ahd_scb_active_in_fifo(struct ahd_softc *ahd, struct scb *scb) |
| { |
| |
| /* |
| * The FIFO is only active for our transaction if |
| * the SCBPTR matches the SCB's ID and the firmware |
| * has installed a handler for the FIFO or we have |
| * a pending SAVEPTRS or CFG4DATA interrupt. |
| */ |
| if (ahd_get_scbptr(ahd) != SCB_GET_TAG(scb) |
| || ((ahd_inb(ahd, LONGJMP_ADDR+1) & INVALID_ADDR) != 0 |
| && (ahd_inb(ahd, SEQINTSRC) & (CFG4DATA|SAVEPTRS)) == 0)) |
| return (0); |
| |
| return (1); |
| } |
| |
| /* |
| * Run a data fifo to completion for a transaction we know |
| * has completed across the SCSI bus (good status has been |
| * received). We are already set to the correct FIFO mode |
| * on entry to this routine. |
| * |
| * This function attempts to operate exactly as the firmware |
| * would when running this FIFO. Care must be taken to update |
| * this routine any time the firmware's FIFO algorithm is |
| * changed. |
| */ |
| static void |
| ahd_run_data_fifo(struct ahd_softc *ahd, struct scb *scb) |
| { |
| u_int seqintsrc; |
| |
| seqintsrc = ahd_inb(ahd, SEQINTSRC); |
| if ((seqintsrc & CFG4DATA) != 0) { |
| uint32_t datacnt; |
| uint32_t sgptr; |
| |
| /* |
| * Clear full residual flag. |
| */ |
| sgptr = ahd_inl_scbram(ahd, SCB_SGPTR) & ~SG_FULL_RESID; |
| ahd_outb(ahd, SCB_SGPTR, sgptr); |
| |
| /* |
| * Load datacnt and address. |
| */ |
| datacnt = ahd_inl_scbram(ahd, SCB_DATACNT); |
| if ((datacnt & AHD_DMA_LAST_SEG) != 0) { |
| sgptr |= LAST_SEG; |
| ahd_outb(ahd, SG_STATE, 0); |
| } else |
| ahd_outb(ahd, SG_STATE, LOADING_NEEDED); |
| ahd_outq(ahd, HADDR, ahd_inq_scbram(ahd, SCB_DATAPTR)); |
| ahd_outl(ahd, HCNT, datacnt & AHD_SG_LEN_MASK); |
| ahd_outb(ahd, SG_CACHE_PRE, sgptr); |
| ahd_outb(ahd, DFCNTRL, PRELOADEN|SCSIEN|HDMAEN); |
| |
| /* |
| * Initialize Residual Fields. |
| */ |
| ahd_outb(ahd, SCB_RESIDUAL_DATACNT+3, datacnt >> 24); |
| ahd_outl(ahd, SCB_RESIDUAL_SGPTR, sgptr & SG_PTR_MASK); |
| |
| /* |
| * Mark the SCB as having a FIFO in use. |
| */ |
| ahd_outb(ahd, SCB_FIFO_USE_COUNT, |
| ahd_inb_scbram(ahd, SCB_FIFO_USE_COUNT) + 1); |
| |
| /* |
| * Install a "fake" handler for this FIFO. |
| */ |
| ahd_outw(ahd, LONGJMP_ADDR, 0); |
| |
| /* |
| * Notify the hardware that we have satisfied |
| * this sequencer interrupt. |
| */ |
| ahd_outb(ahd, CLRSEQINTSRC, CLRCFG4DATA); |
| } else if ((seqintsrc & SAVEPTRS) != 0) { |
| uint32_t sgptr; |
| uint32_t resid; |
| |
| if ((ahd_inb(ahd, LONGJMP_ADDR+1)&INVALID_ADDR) != 0) { |
| /* |
| * Snapshot Save Pointers. All that |
| * is necessary to clear the snapshot |
| * is a CLRCHN. |
| */ |
| goto clrchn; |
| } |
| |
| /* |
| * Disable S/G fetch so the DMA engine |
| * is available to future users. |
| */ |
| if ((ahd_inb(ahd, SG_STATE) & FETCH_INPROG) != 0) |
| ahd_outb(ahd, CCSGCTL, 0); |
| ahd_outb(ahd, SG_STATE, 0); |
| |
| /* |
| * Flush the data FIFO. Strickly only |
| * necessary for Rev A parts. |
| */ |
| ahd_outb(ahd, DFCNTRL, ahd_inb(ahd, DFCNTRL) | FIFOFLUSH); |
| |
| /* |
| * Calculate residual. |
| */ |
| sgptr = ahd_inl_scbram(ahd, SCB_RESIDUAL_SGPTR); |
| resid = ahd_inl(ahd, SHCNT); |
| resid |= ahd_inb_scbram(ahd, SCB_RESIDUAL_DATACNT+3) << 24; |
| ahd_outl(ahd, SCB_RESIDUAL_DATACNT, resid); |
| if ((ahd_inb(ahd, SG_CACHE_SHADOW) & LAST_SEG) == 0) { |
| /* |
| * Must back up to the correct S/G element. |
| * Typically this just means resetting our |
| * low byte to the offset in the SG_CACHE, |
| * but if we wrapped, we have to correct |
| * the other bytes of the sgptr too. |
| */ |
| if ((ahd_inb(ahd, SG_CACHE_SHADOW) & 0x80) != 0 |
| && (sgptr & 0x80) == 0) |
| sgptr -= 0x100; |
| sgptr &= ~0xFF; |
| sgptr |= ahd_inb(ahd, SG_CACHE_SHADOW) |
| & SG_ADDR_MASK; |
| ahd_outl(ahd, SCB_RESIDUAL_SGPTR, sgptr); |
| ahd_outb(ahd, SCB_RESIDUAL_DATACNT + 3, 0); |
| } else if ((resid & AHD_SG_LEN_MASK) == 0) { |
| ahd_outb(ahd, SCB_RESIDUAL_SGPTR, |
| sgptr | SG_LIST_NULL); |
| } |
| /* |
| * Save Pointers. |
| */ |
| ahd_outq(ahd, SCB_DATAPTR, ahd_inq(ahd, SHADDR)); |
| ahd_outl(ahd, SCB_DATACNT, resid); |
| ahd_outl(ahd, SCB_SGPTR, sgptr); |
| ahd_outb(ahd, CLRSEQINTSRC, CLRSAVEPTRS); |
| ahd_outb(ahd, SEQIMODE, |
| ahd_inb(ahd, SEQIMODE) | ENSAVEPTRS); |
| /* |
| * If the data is to the SCSI bus, we are |
| * done, otherwise wait for FIFOEMP. |
| */ |
| if ((ahd_inb(ahd, DFCNTRL) & DIRECTION) != 0) |
| goto clrchn; |
| } else if ((ahd_inb(ahd, SG_STATE) & LOADING_NEEDED) != 0) { |
| uint32_t sgptr; |
| uint64_t data_addr; |
| uint32_t data_len; |
| u_int dfcntrl; |
| |
| /* |
| * Disable S/G fetch so the DMA engine |
| * is available to future users. We won't |
| * be using the DMA engine to load segments. |
| */ |
| if ((ahd_inb(ahd, SG_STATE) & FETCH_INPROG) != 0) { |
| ahd_outb(ahd, CCSGCTL, 0); |
| ahd_outb(ahd, SG_STATE, LOADING_NEEDED); |
| } |
| |
| /* |
| * Wait for the DMA engine to notice that the |
| * host transfer is enabled and that there is |
| * space in the S/G FIFO for new segments before |
| * loading more segments. |
| */ |
| if ((ahd_inb(ahd, DFSTATUS) & PRELOAD_AVAIL) != 0 |
| && (ahd_inb(ahd, DFCNTRL) & HDMAENACK) != 0) { |
| |
| /* |
| * Determine the offset of the next S/G |
| * element to load. |
| */ |
| sgptr = ahd_inl_scbram(ahd, SCB_RESIDUAL_SGPTR); |
| sgptr &= SG_PTR_MASK; |
| if ((ahd->flags & AHD_64BIT_ADDRESSING) != 0) { |
| struct ahd_dma64_seg *sg; |
| |
| sg = ahd_sg_bus_to_virt(ahd, scb, sgptr); |
| data_addr = sg->addr; |
| data_len = sg->len; |
| sgptr += sizeof(*sg); |
| } else { |
| struct ahd_dma_seg *sg; |
| |
| sg = ahd_sg_bus_to_virt(ahd, scb, sgptr); |
| data_addr = sg->len & AHD_SG_HIGH_ADDR_MASK; |
| data_addr <<= 8; |
| data_addr |= sg->addr; |
| data_len = sg->len; |
| sgptr += sizeof(*sg); |
| } |
| |
| /* |
| * Update residual information. |
| */ |
| ahd_outb(ahd, SCB_RESIDUAL_DATACNT+3, data_len >> 24); |
| ahd_outl(ahd, SCB_RESIDUAL_SGPTR, sgptr); |
| |
| /* |
| * Load the S/G. |
| */ |
| if (data_len & AHD_DMA_LAST_SEG) { |
| sgptr |= LAST_SEG; |
| ahd_outb(ahd, SG_STATE, 0); |
| } |
| ahd_outq(ahd, HADDR, data_addr); |
| ahd_outl(ahd, HCNT, data_len & AHD_SG_LEN_MASK); |
| ahd_outb(ahd, SG_CACHE_PRE, sgptr & 0xFF); |
| |
| /* |
| * Advertise the segment to the hardware. |
| */ |
| dfcntrl = ahd_inb(ahd, DFCNTRL)|PRELOADEN|HDMAEN; |
| if ((ahd->features & AHD_NEW_DFCNTRL_OPTS) != 0) { |
| /* |
| * Use SCSIENWRDIS so that SCSIEN |
| * is never modified by this |
| * operation. |
| */ |
| dfcntrl |= SCSIENWRDIS; |
| } |
| ahd_outb(ahd, DFCNTRL, dfcntrl); |
| } |
| } else if ((ahd_inb(ahd, SG_CACHE_SHADOW) & LAST_SEG_DONE) != 0) { |
| |
| /* |
| * Transfer completed to the end of SG list |
| * and has flushed to the host. |
| */ |
| ahd_outb(ahd, SCB_SGPTR, |
| ahd_inb_scbram(ahd, SCB_SGPTR) | SG_LIST_NULL); |
| goto clrchn; |
| } else if ((ahd_inb(ahd, DFSTATUS) & FIFOEMP) != 0) { |
| clrchn: |
| /* |
| * Clear any handler for this FIFO, decrement |
| * the FIFO use count for the SCB, and release |
| * the FIFO. |
| */ |
| ahd_outb(ahd, LONGJMP_ADDR + 1, INVALID_ADDR); |
| ahd_outb(ahd, SCB_FIFO_USE_COUNT, |
| ahd_inb_scbram(ahd, SCB_FIFO_USE_COUNT) - 1); |
| ahd_outb(ahd, DFFSXFRCTL, CLRCHN); |
| } |
| } |
| |
| /* |
| * Look for entries in the QoutFIFO that have completed. |
| * The valid_tag completion field indicates the validity |
| * of the entry - the valid value toggles each time through |
| * the queue. We use the sg_status field in the completion |
| * entry to avoid referencing the hscb if the completion |
| * occurred with no errors and no residual. sg_status is |
| * a copy of the first byte (little endian) of the sgptr |
| * hscb field. |
| */ |
| static void |
| ahd_run_qoutfifo(struct ahd_softc *ahd) |
| { |
| struct ahd_completion *completion; |
| struct scb *scb; |
| u_int scb_index; |
| |
| if ((ahd->flags & AHD_RUNNING_QOUTFIFO) != 0) |
| panic("ahd_run_qoutfifo recursion"); |
| ahd->flags |= AHD_RUNNING_QOUTFIFO; |
| ahd_sync_qoutfifo(ahd, BUS_DMASYNC_POSTREAD); |
| for (;;) { |
| completion = &ahd->qoutfifo[ahd->qoutfifonext]; |
| |
| if (completion->valid_tag != ahd->qoutfifonext_valid_tag) |
| break; |
| |
| scb_index = ahd_le16toh(completion->tag); |
| scb = ahd_lookup_scb(ahd, scb_index); |
| if (scb == NULL) { |
| printk("%s: WARNING no command for scb %d " |
| "(cmdcmplt)\nQOUTPOS = %d\n", |
| ahd_name(ahd), scb_index, |
| ahd->qoutfifonext); |
| ahd_dump_card_state(ahd); |
| } else if ((completion->sg_status & SG_STATUS_VALID) != 0) { |
| ahd_handle_scb_status(ahd, scb); |
| } else { |
| ahd_done(ahd, scb); |
| } |
| |
| ahd->qoutfifonext = (ahd->qoutfifonext+1) & (AHD_QOUT_SIZE-1); |
| if (ahd->qoutfifonext == 0) |
| ahd->qoutfifonext_valid_tag ^= QOUTFIFO_ENTRY_VALID; |
| } |
| ahd->flags &= ~AHD_RUNNING_QOUTFIFO; |
| } |
| |
| /************************* Interrupt Handling *********************************/ |
| static void |
| ahd_handle_hwerrint(struct ahd_softc *ahd) |
| { |
| /* |
| * Some catastrophic hardware error has occurred. |
| * Print it for the user and disable the controller. |
| */ |
| int i; |
| int error; |
| |
| error = ahd_inb(ahd, ERROR); |
| for (i = 0; i < num_errors; i++) { |
| if ((error & ahd_hard_errors[i].errno) != 0) |
| printk("%s: hwerrint, %s\n", |
| ahd_name(ahd), ahd_hard_errors[i].errmesg); |
| } |
| |
| ahd_dump_card_state(ahd); |
| panic("BRKADRINT"); |
| |
| /* Tell everyone that this HBA is no longer available */ |
| ahd_abort_scbs(ahd, CAM_TARGET_WILDCARD, ALL_CHANNELS, |
| CAM_LUN_WILDCARD, SCB_LIST_NULL, ROLE_UNKNOWN, |
| CAM_NO_HBA); |
| |
| /* Tell the system that this controller has gone away. */ |
| ahd_free(ahd); |
| } |
| |
| #ifdef AHD_DEBUG |
| static void |
| ahd_dump_sglist(struct scb *scb) |
| { |
| int i; |
| |
| if (scb->sg_count > 0) { |
| if ((scb->ahd_softc->flags & AHD_64BIT_ADDRESSING) != 0) { |
| struct ahd_dma64_seg *sg_list; |
| |
| sg_list = (struct ahd_dma64_seg*)scb->sg_list; |
| for (i = 0; i < scb->sg_count; i++) { |
| uint64_t addr; |
| uint32_t len; |
| |
| addr = ahd_le64toh(sg_list[i].addr); |
| len = ahd_le32toh(sg_list[i].len); |
| printk("sg[%d] - Addr 0x%x%x : Length %d%s\n", |
| i, |
| (uint32_t)((addr >> 32) & 0xFFFFFFFF), |
| (uint32_t)(addr & 0xFFFFFFFF), |
| sg_list[i].len & AHD_SG_LEN_MASK, |
| (sg_list[i].len & AHD_DMA_LAST_SEG) |
| ? " Last" : ""); |
| } |
| } else { |
| struct ahd_dma_seg *sg_list; |
| |
| sg_list = (struct ahd_dma_seg*)scb->sg_list; |
| for (i = 0; i < scb->sg_count; i++) { |
| uint32_t len; |
| |
| len = ahd_le32toh(sg_list[i].len); |
| printk("sg[%d] - Addr 0x%x%x : Length %d%s\n", |
| i, |
| (len & AHD_SG_HIGH_ADDR_MASK) >> 24, |
| ahd_le32toh(sg_list[i].addr), |
| len & AHD_SG_LEN_MASK, |
| len & AHD_DMA_LAST_SEG ? " Last" : ""); |
| } |
| } |
| } |
| } |
| #endif /* AHD_DEBUG */ |
| |
| static void |
| ahd_handle_seqint(struct ahd_softc *ahd, u_int intstat) |
| { |
| u_int seqintcode; |
| |
| /* |
| * Save the sequencer interrupt code and clear the SEQINT |
| * bit. We will unpause the sequencer, if appropriate, |
| * after servicing the request. |
| */ |
| seqintcode = ahd_inb(ahd, SEQINTCODE); |
| ahd_outb(ahd, CLRINT, CLRSEQINT); |
| if ((ahd->bugs & AHD_INTCOLLISION_BUG) != 0) { |
| /* |
| * Unpause the sequencer and let it clear |
| * SEQINT by writing NO_SEQINT to it. This |
| * will cause the sequencer to be paused again, |
| * which is the expected state of this routine. |
| */ |
| ahd_unpause(ahd); |
| while (!ahd_is_paused(ahd)) |
| ; |
| ahd_outb(ahd, CLRINT, CLRSEQINT); |
| } |
| ahd_update_modes(ahd); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MISC) != 0) |
| printk("%s: Handle Seqint Called for code %d\n", |
| ahd_name(ahd), seqintcode); |
| #endif |
| switch (seqintcode) { |
| case ENTERING_NONPACK: |
| { |
| struct scb *scb; |
| u_int scbid; |
| |
| AHD_ASSERT_MODES(ahd, ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK), |
| ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK)); |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| /* |
| * Somehow need to know if this |
| * is from a selection or reselection. |
| * From that, we can determine target |
| * ID so we at least have an I_T nexus. |
| */ |
| } else { |
| ahd_outb(ahd, SAVED_SCSIID, scb->hscb->scsiid); |
| ahd_outb(ahd, SAVED_LUN, scb->hscb->lun); |
| ahd_outb(ahd, SEQ_FLAGS, 0x0); |
| } |
| if ((ahd_inb(ahd, LQISTAT2) & LQIPHASE_OUTPKT) != 0 |
| && (ahd_inb(ahd, SCSISIGO) & ATNO) != 0) { |
| /* |
| * Phase change after read stream with |
| * CRC error with P0 asserted on last |
| * packet. |
| */ |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_RECOVERY) != 0) |
| printk("%s: Assuming LQIPHASE_NLQ with " |
| "P0 assertion\n", ahd_name(ahd)); |
| #endif |
| } |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_RECOVERY) != 0) |
| printk("%s: Entering NONPACK\n", ahd_name(ahd)); |
| #endif |
| break; |
| } |
| case INVALID_SEQINT: |
| printk("%s: Invalid Sequencer interrupt occurred, " |
| "resetting channel.\n", |
| ahd_name(ahd)); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_RECOVERY) != 0) |
| ahd_dump_card_state(ahd); |
| #endif |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/TRUE); |
| break; |
| case STATUS_OVERRUN: |
| { |
| struct scb *scb; |
| u_int scbid; |
| |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb != NULL) |
| ahd_print_path(ahd, scb); |
| else |
| printk("%s: ", ahd_name(ahd)); |
| printk("SCB %d Packetized Status Overrun", scbid); |
| ahd_dump_card_state(ahd); |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/TRUE); |
| break; |
| } |
| case CFG4ISTAT_INTR: |
| { |
| struct scb *scb; |
| u_int scbid; |
| |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| ahd_dump_card_state(ahd); |
| printk("CFG4ISTAT: Free SCB %d referenced", scbid); |
| panic("For safety"); |
| } |
| ahd_outq(ahd, HADDR, scb->sense_busaddr); |
| ahd_outw(ahd, HCNT, AHD_SENSE_BUFSIZE); |
| ahd_outb(ahd, HCNT + 2, 0); |
| ahd_outb(ahd, SG_CACHE_PRE, SG_LAST_SEG); |
| ahd_outb(ahd, DFCNTRL, PRELOADEN|SCSIEN|HDMAEN); |
| break; |
| } |
| case ILLEGAL_PHASE: |
| { |
| u_int bus_phase; |
| |
| bus_phase = ahd_inb(ahd, SCSISIGI) & PHASE_MASK; |
| printk("%s: ILLEGAL_PHASE 0x%x\n", |
| ahd_name(ahd), bus_phase); |
| |
| switch (bus_phase) { |
| case P_DATAOUT: |
| case P_DATAIN: |
| case P_DATAOUT_DT: |
| case P_DATAIN_DT: |
| case P_MESGOUT: |
| case P_STATUS: |
| case P_MESGIN: |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/TRUE); |
| printk("%s: Issued Bus Reset.\n", ahd_name(ahd)); |
| break; |
| case P_COMMAND: |
| { |
| struct ahd_devinfo devinfo; |
| struct scb *scb; |
| struct ahd_initiator_tinfo *targ_info; |
| struct ahd_tmode_tstate *tstate; |
| struct ahd_transinfo *tinfo; |
| u_int scbid; |
| |
| /* |
| * If a target takes us into the command phase |
| * assume that it has been externally reset and |
| * has thus lost our previous packetized negotiation |
| * agreement. Since we have not sent an identify |
| * message and may not have fully qualified the |
| * connection, we change our command to TUR, assert |
| * ATN and ABORT the task when we go to message in |
| * phase. The OSM will see the REQUEUE_REQUEST |
| * status and retry the command. |
| */ |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| printk("Invalid phase with no valid SCB. " |
| "Resetting bus.\n"); |
| ahd_reset_channel(ahd, 'A', |
| /*Initiate Reset*/TRUE); |
| break; |
| } |
| ahd_compile_devinfo(&devinfo, SCB_GET_OUR_ID(scb), |
| SCB_GET_TARGET(ahd, scb), |
| SCB_GET_LUN(scb), |
| SCB_GET_CHANNEL(ahd, scb), |
| ROLE_INITIATOR); |
| targ_info = ahd_fetch_transinfo(ahd, |
| devinfo.channel, |
| devinfo.our_scsiid, |
| devinfo.target, |
| &tstate); |
| tinfo = &targ_info->curr; |
| ahd_set_width(ahd, &devinfo, MSG_EXT_WDTR_BUS_8_BIT, |
| AHD_TRANS_ACTIVE, /*paused*/TRUE); |
| ahd_set_syncrate(ahd, &devinfo, /*period*/0, |
| /*offset*/0, /*ppr_options*/0, |
| AHD_TRANS_ACTIVE, /*paused*/TRUE); |
| /* Hand-craft TUR command */ |
| ahd_outb(ahd, SCB_CDB_STORE, 0); |
| ahd_outb(ahd, SCB_CDB_STORE+1, 0); |
| ahd_outb(ahd, SCB_CDB_STORE+2, 0); |
| ahd_outb(ahd, SCB_CDB_STORE+3, 0); |
| ahd_outb(ahd, SCB_CDB_STORE+4, 0); |
| ahd_outb(ahd, SCB_CDB_STORE+5, 0); |
| ahd_outb(ahd, SCB_CDB_LEN, 6); |
| scb->hscb->control &= ~(TAG_ENB|SCB_TAG_TYPE); |
| scb->hscb->control |= MK_MESSAGE; |
| ahd_outb(ahd, SCB_CONTROL, scb->hscb->control); |
| ahd_outb(ahd, MSG_OUT, HOST_MSG); |
| ahd_outb(ahd, SAVED_SCSIID, scb->hscb->scsiid); |
| /* |
| * The lun is 0, regardless of the SCB's lun |
| * as we have not sent an identify message. |
| */ |
| ahd_outb(ahd, SAVED_LUN, 0); |
| ahd_outb(ahd, SEQ_FLAGS, 0); |
| ahd_assert_atn(ahd); |
| scb->flags &= ~SCB_PACKETIZED; |
| scb->flags |= SCB_ABORT|SCB_EXTERNAL_RESET; |
| ahd_freeze_devq(ahd, scb); |
| ahd_set_transaction_status(scb, CAM_REQUEUE_REQ); |
| ahd_freeze_scb(scb); |
| |
| /* Notify XPT */ |
| ahd_send_async(ahd, devinfo.channel, devinfo.target, |
| CAM_LUN_WILDCARD, AC_SENT_BDR); |
| |
| /* |
| * Allow the sequencer to continue with |
| * non-pack processing. |
| */ |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| ahd_outb(ahd, CLRLQOINT1, CLRLQOPHACHGINPKT); |
| if ((ahd->bugs & AHD_CLRLQO_AUTOCLR_BUG) != 0) { |
| ahd_outb(ahd, CLRLQOINT1, 0); |
| } |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_RECOVERY) != 0) { |
| ahd_print_path(ahd, scb); |
| printk("Unexpected command phase from " |
| "packetized target\n"); |
| } |
| #endif |
| break; |
| } |
| } |
| break; |
| } |
| case CFG4OVERRUN: |
| { |
| struct scb *scb; |
| u_int scb_index; |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_RECOVERY) != 0) { |
| printk("%s: CFG4OVERRUN mode = %x\n", ahd_name(ahd), |
| ahd_inb(ahd, MODE_PTR)); |
| } |
| #endif |
| scb_index = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scb_index); |
| if (scb == NULL) { |
| /* |
| * Attempt to transfer to an SCB that is |
| * not outstanding. |
| */ |
| ahd_assert_atn(ahd); |
| ahd_outb(ahd, MSG_OUT, HOST_MSG); |
| ahd->msgout_buf[0] = MSG_ABORT_TASK; |
| ahd->msgout_len = 1; |
| ahd->msgout_index = 0; |
| ahd->msg_type = MSG_TYPE_INITIATOR_MSGOUT; |
| /* |
| * Clear status received flag to prevent any |
| * attempt to complete this bogus SCB. |
| */ |
| ahd_outb(ahd, SCB_CONTROL, |
| ahd_inb_scbram(ahd, SCB_CONTROL) |
| & ~STATUS_RCVD); |
| } |
| break; |
| } |
| case DUMP_CARD_STATE: |
| { |
| ahd_dump_card_state(ahd); |
| break; |
| } |
| case PDATA_REINIT: |
| { |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_RECOVERY) != 0) { |
| printk("%s: PDATA_REINIT - DFCNTRL = 0x%x " |
| "SG_CACHE_SHADOW = 0x%x\n", |
| ahd_name(ahd), ahd_inb(ahd, DFCNTRL), |
| ahd_inb(ahd, SG_CACHE_SHADOW)); |
| } |
| #endif |
| ahd_reinitialize_dataptrs(ahd); |
| break; |
| } |
| case HOST_MSG_LOOP: |
| { |
| struct ahd_devinfo devinfo; |
| |
| /* |
| * The sequencer has encountered a message phase |
| * that requires host assistance for completion. |
| * While handling the message phase(s), we will be |
| * notified by the sequencer after each byte is |
| * transferred so we can track bus phase changes. |
| * |
| * If this is the first time we've seen a HOST_MSG_LOOP |
| * interrupt, initialize the state of the host message |
| * loop. |
| */ |
| ahd_fetch_devinfo(ahd, &devinfo); |
| if (ahd->msg_type == MSG_TYPE_NONE) { |
| struct scb *scb; |
| u_int scb_index; |
| u_int bus_phase; |
| |
| bus_phase = ahd_inb(ahd, SCSISIGI) & PHASE_MASK; |
| if (bus_phase != P_MESGIN |
| && bus_phase != P_MESGOUT) { |
| printk("ahd_intr: HOST_MSG_LOOP bad " |
| "phase 0x%x\n", bus_phase); |
| /* |
| * Probably transitioned to bus free before |
| * we got here. Just punt the message. |
| */ |
| ahd_dump_card_state(ahd); |
| ahd_clear_intstat(ahd); |
| ahd_restart(ahd); |
| return; |
| } |
| |
| scb_index = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scb_index); |
| if (devinfo.role == ROLE_INITIATOR) { |
| if (bus_phase == P_MESGOUT) |
| ahd_setup_initiator_msgout(ahd, |
| &devinfo, |
| scb); |
| else { |
| ahd->msg_type = |
| MSG_TYPE_INITIATOR_MSGIN; |
| ahd->msgin_index = 0; |
| } |
| } |
| #ifdef AHD_TARGET_MODE |
| else { |
| if (bus_phase == P_MESGOUT) { |
| ahd->msg_type = |
| MSG_TYPE_TARGET_MSGOUT; |
| ahd->msgin_index = 0; |
| } |
| else |
| ahd_setup_target_msgin(ahd, |
| &devinfo, |
| scb); |
| } |
| #endif |
| } |
| |
| ahd_handle_message_phase(ahd); |
| break; |
| } |
| case NO_MATCH: |
| { |
| /* Ensure we don't leave the selection hardware on */ |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| ahd_outb(ahd, SCSISEQ0, ahd_inb(ahd, SCSISEQ0) & ~ENSELO); |
| |
| printk("%s:%c:%d: no active SCB for reconnecting " |
| "target - issuing BUS DEVICE RESET\n", |
| ahd_name(ahd), 'A', ahd_inb(ahd, SELID) >> 4); |
| printk("SAVED_SCSIID == 0x%x, SAVED_LUN == 0x%x, " |
| "REG0 == 0x%x ACCUM = 0x%x\n", |
| ahd_inb(ahd, SAVED_SCSIID), ahd_inb(ahd, SAVED_LUN), |
| ahd_inw(ahd, REG0), ahd_inb(ahd, ACCUM)); |
| printk("SEQ_FLAGS == 0x%x, SCBPTR == 0x%x, BTT == 0x%x, " |
| "SINDEX == 0x%x\n", |
| ahd_inb(ahd, SEQ_FLAGS), ahd_get_scbptr(ahd), |
| ahd_find_busy_tcl(ahd, |
| BUILD_TCL(ahd_inb(ahd, SAVED_SCSIID), |
| ahd_inb(ahd, SAVED_LUN))), |
| ahd_inw(ahd, SINDEX)); |
| printk("SELID == 0x%x, SCB_SCSIID == 0x%x, SCB_LUN == 0x%x, " |
| "SCB_CONTROL == 0x%x\n", |
| ahd_inb(ahd, SELID), ahd_inb_scbram(ahd, SCB_SCSIID), |
| ahd_inb_scbram(ahd, SCB_LUN), |
| ahd_inb_scbram(ahd, SCB_CONTROL)); |
| printk("SCSIBUS[0] == 0x%x, SCSISIGI == 0x%x\n", |
| ahd_inb(ahd, SCSIBUS), ahd_inb(ahd, SCSISIGI)); |
| printk("SXFRCTL0 == 0x%x\n", ahd_inb(ahd, SXFRCTL0)); |
| printk("SEQCTL0 == 0x%x\n", ahd_inb(ahd, SEQCTL0)); |
| ahd_dump_card_state(ahd); |
| ahd->msgout_buf[0] = MSG_BUS_DEV_RESET; |
| ahd->msgout_len = 1; |
| ahd->msgout_index = 0; |
| ahd->msg_type = MSG_TYPE_INITIATOR_MSGOUT; |
| ahd_outb(ahd, MSG_OUT, HOST_MSG); |
| ahd_assert_atn(ahd); |
| break; |
| } |
| case PROTO_VIOLATION: |
| { |
| ahd_handle_proto_violation(ahd); |
| break; |
| } |
| case IGN_WIDE_RES: |
| { |
| struct ahd_devinfo devinfo; |
| |
| ahd_fetch_devinfo(ahd, &devinfo); |
| ahd_handle_ign_wide_residue(ahd, &devinfo); |
| break; |
| } |
| case BAD_PHASE: |
| { |
| u_int lastphase; |
| |
| lastphase = ahd_inb(ahd, LASTPHASE); |
| printk("%s:%c:%d: unknown scsi bus phase %x, " |
| "lastphase = 0x%x. Attempting to continue\n", |
| ahd_name(ahd), 'A', |
| SCSIID_TARGET(ahd, ahd_inb(ahd, SAVED_SCSIID)), |
| lastphase, ahd_inb(ahd, SCSISIGI)); |
| break; |
| } |
| case MISSED_BUSFREE: |
| { |
| u_int lastphase; |
| |
| lastphase = ahd_inb(ahd, LASTPHASE); |
| printk("%s:%c:%d: Missed busfree. " |
| "Lastphase = 0x%x, Curphase = 0x%x\n", |
| ahd_name(ahd), 'A', |
| SCSIID_TARGET(ahd, ahd_inb(ahd, SAVED_SCSIID)), |
| lastphase, ahd_inb(ahd, SCSISIGI)); |
| ahd_restart(ahd); |
| return; |
| } |
| case DATA_OVERRUN: |
| { |
| /* |
| * When the sequencer detects an overrun, it |
| * places the controller in "BITBUCKET" mode |
| * and allows the target to complete its transfer. |
| * Unfortunately, none of the counters get updated |
| * when the controller is in this mode, so we have |
| * no way of knowing how large the overrun was. |
| */ |
| struct scb *scb; |
| u_int scbindex; |
| #ifdef AHD_DEBUG |
| u_int lastphase; |
| #endif |
| |
| scbindex = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbindex); |
| #ifdef AHD_DEBUG |
| lastphase = ahd_inb(ahd, LASTPHASE); |
| if ((ahd_debug & AHD_SHOW_RECOVERY) != 0) { |
| ahd_print_path(ahd, scb); |
| printk("data overrun detected %s. Tag == 0x%x.\n", |
| ahd_lookup_phase_entry(lastphase)->phasemsg, |
| SCB_GET_TAG(scb)); |
| ahd_print_path(ahd, scb); |
| printk("%s seen Data Phase. Length = %ld. " |
| "NumSGs = %d.\n", |
| ahd_inb(ahd, SEQ_FLAGS) & DPHASE |
| ? "Have" : "Haven't", |
| ahd_get_transfer_length(scb), scb->sg_count); |
| ahd_dump_sglist(scb); |
| } |
| #endif |
| |
| /* |
| * Set this and it will take effect when the |
| * target does a command complete. |
| */ |
| ahd_freeze_devq(ahd, scb); |
| ahd_set_transaction_status(scb, CAM_DATA_RUN_ERR); |
| ahd_freeze_scb(scb); |
| break; |
| } |
| case MKMSG_FAILED: |
| { |
| struct ahd_devinfo devinfo; |
| struct scb *scb; |
| u_int scbid; |
| |
| ahd_fetch_devinfo(ahd, &devinfo); |
| printk("%s:%c:%d:%d: Attempt to issue message failed\n", |
| ahd_name(ahd), devinfo.channel, devinfo.target, |
| devinfo.lun); |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb != NULL |
| && (scb->flags & SCB_RECOVERY_SCB) != 0) |
| /* |
| * Ensure that we didn't put a second instance of this |
| * SCB into the QINFIFO. |
| */ |
| ahd_search_qinfifo(ahd, SCB_GET_TARGET(ahd, scb), |
| SCB_GET_CHANNEL(ahd, scb), |
| SCB_GET_LUN(scb), SCB_GET_TAG(scb), |
| ROLE_INITIATOR, /*status*/0, |
| SEARCH_REMOVE); |
| ahd_outb(ahd, SCB_CONTROL, |
| ahd_inb_scbram(ahd, SCB_CONTROL) & ~MK_MESSAGE); |
| break; |
| } |
| case TASKMGMT_FUNC_COMPLETE: |
| { |
| u_int scbid; |
| struct scb *scb; |
| |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb != NULL) { |
| u_int lun; |
| u_int tag; |
| cam_status error; |
| |
| ahd_print_path(ahd, scb); |
| printk("Task Management Func 0x%x Complete\n", |
| scb->hscb->task_management); |
| lun = CAM_LUN_WILDCARD; |
| tag = SCB_LIST_NULL; |
| |
| switch (scb->hscb->task_management) { |
| case SIU_TASKMGMT_ABORT_TASK: |
| tag = SCB_GET_TAG(scb); |
| case SIU_TASKMGMT_ABORT_TASK_SET: |
| case SIU_TASKMGMT_CLEAR_TASK_SET: |
| lun = scb->hscb->lun; |
| error = CAM_REQ_ABORTED; |
| ahd_abort_scbs(ahd, SCB_GET_TARGET(ahd, scb), |
| 'A', lun, tag, ROLE_INITIATOR, |
| error); |
| break; |
| case SIU_TASKMGMT_LUN_RESET: |
| lun = scb->hscb->lun; |
| case SIU_TASKMGMT_TARGET_RESET: |
| { |
| struct ahd_devinfo devinfo; |
| |
| ahd_scb_devinfo(ahd, &devinfo, scb); |
| error = CAM_BDR_SENT; |
| ahd_handle_devreset(ahd, &devinfo, lun, |
| CAM_BDR_SENT, |
| lun != CAM_LUN_WILDCARD |
| ? "Lun Reset" |
| : "Target Reset", |
| /*verbose_level*/0); |
| break; |
| } |
| default: |
| panic("Unexpected TaskMgmt Func\n"); |
| break; |
| } |
| } |
| break; |
| } |
| case TASKMGMT_CMD_CMPLT_OKAY: |
| { |
| u_int scbid; |
| struct scb *scb; |
| |
| /* |
| * An ABORT TASK TMF failed to be delivered before |
| * the targeted command completed normally. |
| */ |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb != NULL) { |
| /* |
| * Remove the second instance of this SCB from |
| * the QINFIFO if it is still there. |
| */ |
| ahd_print_path(ahd, scb); |
| printk("SCB completes before TMF\n"); |
| /* |
| * Handle losing the race. Wait until any |
| * current selection completes. We will then |
| * set the TMF back to zero in this SCB so that |
| * the sequencer doesn't bother to issue another |
| * sequencer interrupt for its completion. |
| */ |
| while ((ahd_inb(ahd, SCSISEQ0) & ENSELO) != 0 |
| && (ahd_inb(ahd, SSTAT0) & SELDO) == 0 |
| && (ahd_inb(ahd, SSTAT1) & SELTO) == 0) |
| ; |
| ahd_outb(ahd, SCB_TASK_MANAGEMENT, 0); |
| ahd_search_qinfifo(ahd, SCB_GET_TARGET(ahd, scb), |
| SCB_GET_CHANNEL(ahd, scb), |
| SCB_GET_LUN(scb), SCB_GET_TAG(scb), |
| ROLE_INITIATOR, /*status*/0, |
| SEARCH_REMOVE); |
| } |
| break; |
| } |
| case TRACEPOINT0: |
| case TRACEPOINT1: |
| case TRACEPOINT2: |
| case TRACEPOINT3: |
| printk("%s: Tracepoint %d\n", ahd_name(ahd), |
| seqintcode - TRACEPOINT0); |
| break; |
| case NO_SEQINT: |
| break; |
| case SAW_HWERR: |
| ahd_handle_hwerrint(ahd); |
| break; |
| default: |
| printk("%s: Unexpected SEQINTCODE %d\n", ahd_name(ahd), |
| seqintcode); |
| break; |
| } |
| /* |
| * The sequencer is paused immediately on |
| * a SEQINT, so we should restart it when |
| * we're done. |
| */ |
| ahd_unpause(ahd); |
| } |
| |
| static void |
| ahd_handle_scsiint(struct ahd_softc *ahd, u_int intstat) |
| { |
| struct scb *scb; |
| u_int status0; |
| u_int status3; |
| u_int status; |
| u_int lqistat1; |
| u_int lqostat0; |
| u_int scbid; |
| u_int busfreetime; |
| |
| ahd_update_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| |
| status3 = ahd_inb(ahd, SSTAT3) & (NTRAMPERR|OSRAMPERR); |
| status0 = ahd_inb(ahd, SSTAT0) & (IOERR|OVERRUN|SELDI|SELDO); |
| status = ahd_inb(ahd, SSTAT1) & (SELTO|SCSIRSTI|BUSFREE|SCSIPERR); |
| lqistat1 = ahd_inb(ahd, LQISTAT1); |
| lqostat0 = ahd_inb(ahd, LQOSTAT0); |
| busfreetime = ahd_inb(ahd, SSTAT2) & BUSFREETIME; |
| |
| /* |
| * Ignore external resets after a bus reset. |
| */ |
| if (((status & SCSIRSTI) != 0) && (ahd->flags & AHD_BUS_RESET_ACTIVE)) { |
| ahd_outb(ahd, CLRSINT1, CLRSCSIRSTI); |
| return; |
| } |
| |
| /* |
| * Clear bus reset flag |
| */ |
| ahd->flags &= ~AHD_BUS_RESET_ACTIVE; |
| |
| if ((status0 & (SELDI|SELDO)) != 0) { |
| u_int simode0; |
| |
| ahd_set_modes(ahd, AHD_MODE_CFG, AHD_MODE_CFG); |
| simode0 = ahd_inb(ahd, SIMODE0); |
| status0 &= simode0 & (IOERR|OVERRUN|SELDI|SELDO); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| } |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb != NULL |
| && (ahd_inb(ahd, SEQ_FLAGS) & NOT_IDENTIFIED) != 0) |
| scb = NULL; |
| |
| if ((status0 & IOERR) != 0) { |
| u_int now_lvd; |
| |
| now_lvd = ahd_inb(ahd, SBLKCTL) & ENAB40; |
| printk("%s: Transceiver State Has Changed to %s mode\n", |
| ahd_name(ahd), now_lvd ? "LVD" : "SE"); |
| ahd_outb(ahd, CLRSINT0, CLRIOERR); |
| /* |
| * A change in I/O mode is equivalent to a bus reset. |
| */ |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/TRUE); |
| ahd_pause(ahd); |
| ahd_setup_iocell_workaround(ahd); |
| ahd_unpause(ahd); |
| } else if ((status0 & OVERRUN) != 0) { |
| |
| printk("%s: SCSI offset overrun detected. Resetting bus.\n", |
| ahd_name(ahd)); |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/TRUE); |
| } else if ((status & SCSIRSTI) != 0) { |
| |
| printk("%s: Someone reset channel A\n", ahd_name(ahd)); |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/FALSE); |
| } else if ((status & SCSIPERR) != 0) { |
| |
| /* Make sure the sequencer is in a safe location. */ |
| ahd_clear_critical_section(ahd); |
| |
| ahd_handle_transmission_error(ahd); |
| } else if (lqostat0 != 0) { |
| |
| printk("%s: lqostat0 == 0x%x!\n", ahd_name(ahd), lqostat0); |
| ahd_outb(ahd, CLRLQOINT0, lqostat0); |
| if ((ahd->bugs & AHD_CLRLQO_AUTOCLR_BUG) != 0) |
| ahd_outb(ahd, CLRLQOINT1, 0); |
| } else if ((status & SELTO) != 0) { |
| /* Stop the selection */ |
| ahd_outb(ahd, SCSISEQ0, 0); |
| |
| /* Make sure the sequencer is in a safe location. */ |
| ahd_clear_critical_section(ahd); |
| |
| /* No more pending messages */ |
| ahd_clear_msg_state(ahd); |
| |
| /* Clear interrupt state */ |
| ahd_outb(ahd, CLRSINT1, CLRSELTIMEO|CLRBUSFREE|CLRSCSIPERR); |
| |
| /* |
| * Although the driver does not care about the |
| * 'Selection in Progress' status bit, the busy |
| * LED does. SELINGO is only cleared by a successful |
| * selection, so we must manually clear it to insure |
| * the LED turns off just incase no future successful |
| * selections occur (e.g. no devices on the bus). |
| */ |
| ahd_outb(ahd, CLRSINT0, CLRSELINGO); |
| |
| scbid = ahd_inw(ahd, WAITING_TID_HEAD); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| printk("%s: ahd_intr - referenced scb not " |
| "valid during SELTO scb(0x%x)\n", |
| ahd_name(ahd), scbid); |
| ahd_dump_card_state(ahd); |
| } else { |
| struct ahd_devinfo devinfo; |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_SELTO) != 0) { |
| ahd_print_path(ahd, scb); |
| printk("Saw Selection Timeout for SCB 0x%x\n", |
| scbid); |
| } |
| #endif |
| ahd_scb_devinfo(ahd, &devinfo, scb); |
| ahd_set_transaction_status(scb, CAM_SEL_TIMEOUT); |
| ahd_freeze_devq(ahd, scb); |
| |
| /* |
| * Cancel any pending transactions on the device |
| * now that it seems to be missing. This will |
| * also revert us to async/narrow transfers until |
| * we can renegotiate with the device. |
| */ |
| ahd_handle_devreset(ahd, &devinfo, |
| CAM_LUN_WILDCARD, |
| CAM_SEL_TIMEOUT, |
| "Selection Timeout", |
| /*verbose_level*/1); |
| } |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| ahd_iocell_first_selection(ahd); |
| ahd_unpause(ahd); |
| } else if ((status0 & (SELDI|SELDO)) != 0) { |
| |
| ahd_iocell_first_selection(ahd); |
| ahd_unpause(ahd); |
| } else if (status3 != 0) { |
| printk("%s: SCSI Cell parity error SSTAT3 == 0x%x\n", |
| ahd_name(ahd), status3); |
| ahd_outb(ahd, CLRSINT3, status3); |
| } else if ((lqistat1 & (LQIPHASE_LQ|LQIPHASE_NLQ)) != 0) { |
| |
| /* Make sure the sequencer is in a safe location. */ |
| ahd_clear_critical_section(ahd); |
| |
| ahd_handle_lqiphase_error(ahd, lqistat1); |
| } else if ((lqistat1 & LQICRCI_NLQ) != 0) { |
| /* |
| * This status can be delayed during some |
| * streaming operations. The SCSIPHASE |
| * handler has already dealt with this case |
| * so just clear the error. |
| */ |
| ahd_outb(ahd, CLRLQIINT1, CLRLQICRCI_NLQ); |
| } else if ((status & BUSFREE) != 0 |
| || (lqistat1 & LQOBUSFREE) != 0) { |
| u_int lqostat1; |
| int restart; |
| int clear_fifo; |
| int packetized; |
| u_int mode; |
| |
| /* |
| * Clear our selection hardware as soon as possible. |
| * We may have an entry in the waiting Q for this target, |
| * that is affected by this busfree and we don't want to |
| * go about selecting the target while we handle the event. |
| */ |
| ahd_outb(ahd, SCSISEQ0, 0); |
| |
| /* Make sure the sequencer is in a safe location. */ |
| ahd_clear_critical_section(ahd); |
| |
| /* |
| * Determine what we were up to at the time of |
| * the busfree. |
| */ |
| mode = AHD_MODE_SCSI; |
| busfreetime = ahd_inb(ahd, SSTAT2) & BUSFREETIME; |
| lqostat1 = ahd_inb(ahd, LQOSTAT1); |
| switch (busfreetime) { |
| case BUSFREE_DFF0: |
| case BUSFREE_DFF1: |
| { |
| mode = busfreetime == BUSFREE_DFF0 |
| ? AHD_MODE_DFF0 : AHD_MODE_DFF1; |
| ahd_set_modes(ahd, mode, mode); |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| printk("%s: Invalid SCB %d in DFF%d " |
| "during unexpected busfree\n", |
| ahd_name(ahd), scbid, mode); |
| packetized = 0; |
| } else |
| packetized = (scb->flags & SCB_PACKETIZED) != 0; |
| clear_fifo = 1; |
| break; |
| } |
| case BUSFREE_LQO: |
| clear_fifo = 0; |
| packetized = 1; |
| break; |
| default: |
| clear_fifo = 0; |
| packetized = (lqostat1 & LQOBUSFREE) != 0; |
| if (!packetized |
| && ahd_inb(ahd, LASTPHASE) == P_BUSFREE |
| && (ahd_inb(ahd, SSTAT0) & SELDI) == 0 |
| && ((ahd_inb(ahd, SSTAT0) & SELDO) == 0 |
| || (ahd_inb(ahd, SCSISEQ0) & ENSELO) == 0)) |
| /* |
| * Assume packetized if we are not |
| * on the bus in a non-packetized |
| * capacity and any pending selection |
| * was a packetized selection. |
| */ |
| packetized = 1; |
| break; |
| } |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MISC) != 0) |
| printk("Saw Busfree. Busfreetime = 0x%x.\n", |
| busfreetime); |
| #endif |
| /* |
| * Busfrees that occur in non-packetized phases are |
| * handled by the nonpkt_busfree handler. |
| */ |
| if (packetized && ahd_inb(ahd, LASTPHASE) == P_BUSFREE) { |
| restart = ahd_handle_pkt_busfree(ahd, busfreetime); |
| } else { |
| packetized = 0; |
| restart = ahd_handle_nonpkt_busfree(ahd); |
| } |
| /* |
| * Clear the busfree interrupt status. The setting of |
| * the interrupt is a pulse, so in a perfect world, we |
| * would not need to muck with the ENBUSFREE logic. This |
| * would ensure that if the bus moves on to another |
| * connection, busfree protection is still in force. If |
| * BUSFREEREV is broken, however, we must manually clear |
| * the ENBUSFREE if the busfree occurred during a non-pack |
| * connection so that we don't get false positives during |
| * future, packetized, connections. |
| */ |
| ahd_outb(ahd, CLRSINT1, CLRBUSFREE); |
| if (packetized == 0 |
| && (ahd->bugs & AHD_BUSFREEREV_BUG) != 0) |
| ahd_outb(ahd, SIMODE1, |
| ahd_inb(ahd, SIMODE1) & ~ENBUSFREE); |
| |
| if (clear_fifo) |
| ahd_clear_fifo(ahd, mode); |
| |
| ahd_clear_msg_state(ahd); |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| if (restart) { |
| ahd_restart(ahd); |
| } else { |
| ahd_unpause(ahd); |
| } |
| } else { |
| printk("%s: Missing case in ahd_handle_scsiint. status = %x\n", |
| ahd_name(ahd), status); |
| ahd_dump_card_state(ahd); |
| ahd_clear_intstat(ahd); |
| ahd_unpause(ahd); |
| } |
| } |
| |
| static void |
| ahd_handle_transmission_error(struct ahd_softc *ahd) |
| { |
| struct scb *scb; |
| u_int scbid; |
| u_int lqistat1; |
| u_int lqistat2; |
| u_int msg_out; |
| u_int curphase; |
| u_int lastphase; |
| u_int perrdiag; |
| u_int cur_col; |
| int silent; |
| |
| scb = NULL; |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| lqistat1 = ahd_inb(ahd, LQISTAT1) & ~(LQIPHASE_LQ|LQIPHASE_NLQ); |
| lqistat2 = ahd_inb(ahd, LQISTAT2); |
| if ((lqistat1 & (LQICRCI_NLQ|LQICRCI_LQ)) == 0 |
| && (ahd->bugs & AHD_NLQICRC_DELAYED_BUG) != 0) { |
| u_int lqistate; |
| |
| ahd_set_modes(ahd, AHD_MODE_CFG, AHD_MODE_CFG); |
| lqistate = ahd_inb(ahd, LQISTATE); |
| if ((lqistate >= 0x1E && lqistate <= 0x24) |
| || (lqistate == 0x29)) { |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_RECOVERY) != 0) { |
| printk("%s: NLQCRC found via LQISTATE\n", |
| ahd_name(ahd)); |
| } |
| #endif |
| lqistat1 |= LQICRCI_NLQ; |
| } |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| } |
| |
| ahd_outb(ahd, CLRLQIINT1, lqistat1); |
| lastphase = ahd_inb(ahd, LASTPHASE); |
| curphase = ahd_inb(ahd, SCSISIGI) & PHASE_MASK; |
| perrdiag = ahd_inb(ahd, PERRDIAG); |
| msg_out = MSG_INITIATOR_DET_ERR; |
| ahd_outb(ahd, CLRSINT1, CLRSCSIPERR); |
| |
| /* |
| * Try to find the SCB associated with this error. |
| */ |
| silent = FALSE; |
| if (lqistat1 == 0 |
| || (lqistat1 & LQICRCI_NLQ) != 0) { |
| if ((lqistat1 & (LQICRCI_NLQ|LQIOVERI_NLQ)) != 0) |
| ahd_set_active_fifo(ahd); |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb != NULL && SCB_IS_SILENT(scb)) |
| silent = TRUE; |
| } |
| |
| cur_col = 0; |
| if (silent == FALSE) { |
| printk("%s: Transmission error detected\n", ahd_name(ahd)); |
| ahd_lqistat1_print(lqistat1, &cur_col, 50); |
| ahd_lastphase_print(lastphase, &cur_col, 50); |
| ahd_scsisigi_print(curphase, &cur_col, 50); |
| ahd_perrdiag_print(perrdiag, &cur_col, 50); |
| printk("\n"); |
| ahd_dump_card_state(ahd); |
| } |
| |
| if ((lqistat1 & (LQIOVERI_LQ|LQIOVERI_NLQ)) != 0) { |
| if (silent == FALSE) { |
| printk("%s: Gross protocol error during incoming " |
| "packet. lqistat1 == 0x%x. Resetting bus.\n", |
| ahd_name(ahd), lqistat1); |
| } |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/TRUE); |
| return; |
| } else if ((lqistat1 & LQICRCI_LQ) != 0) { |
| /* |
| * A CRC error has been detected on an incoming LQ. |
| * The bus is currently hung on the last ACK. |
| * Hit LQIRETRY to release the last ack, and |
| * wait for the sequencer to determine that ATNO |
| * is asserted while in message out to take us |
| * to our host message loop. No NONPACKREQ or |
| * LQIPHASE type errors will occur in this |
| * scenario. After this first LQIRETRY, the LQI |
| * manager will be in ISELO where it will |
| * happily sit until another packet phase begins. |
| * Unexpected bus free detection is enabled |
| * through any phases that occur after we release |
| * this last ack until the LQI manager sees a |
| * packet phase. This implies we may have to |
| * ignore a perfectly valid "unexected busfree" |
| * after our "initiator detected error" message is |
| * sent. A busfree is the expected response after |
| * we tell the target that it's L_Q was corrupted. |
| * (SPI4R09 10.7.3.3.3) |
| */ |
| ahd_outb(ahd, LQCTL2, LQIRETRY); |
| printk("LQIRetry for LQICRCI_LQ to release ACK\n"); |
| } else if ((lqistat1 & LQICRCI_NLQ) != 0) { |
| /* |
| * We detected a CRC error in a NON-LQ packet. |
| * The hardware has varying behavior in this situation |
| * depending on whether this packet was part of a |
| * stream or not. |
| * |
| * PKT by PKT mode: |
| * The hardware has already acked the complete packet. |
| * If the target honors our outstanding ATN condition, |
| * we should be (or soon will be) in MSGOUT phase. |
| * This will trigger the LQIPHASE_LQ status bit as the |
| * hardware was expecting another LQ. Unexpected |
| * busfree detection is enabled. Once LQIPHASE_LQ is |
| * true (first entry into host message loop is much |
| * the same), we must clear LQIPHASE_LQ and hit |
| * LQIRETRY so the hardware is ready to handle |
| * a future LQ. NONPACKREQ will not be asserted again |
| * once we hit LQIRETRY until another packet is |
| * processed. The target may either go busfree |
| * or start another packet in response to our message. |
| * |
| * Read Streaming P0 asserted: |
| * If we raise ATN and the target completes the entire |
| * stream (P0 asserted during the last packet), the |
| * hardware will ack all data and return to the ISTART |
| * state. When the target reponds to our ATN condition, |
| * LQIPHASE_LQ will be asserted. We should respond to |
| * this with an LQIRETRY to prepare for any future |
| * packets. NONPACKREQ will not be asserted again |
| * once we hit LQIRETRY until another packet is |
| * processed. The target may either go busfree or |
| * start another packet in response to our message. |
| * Busfree detection is enabled. |
| * |
| * Read Streaming P0 not asserted: |
| * If we raise ATN and the target transitions to |
| * MSGOUT in or after a packet where P0 is not |
| * asserted, the hardware will assert LQIPHASE_NLQ. |
| * We should respond to the LQIPHASE_NLQ with an |
| * LQIRETRY. Should the target stay in a non-pkt |
| * phase after we send our message, the hardware |
| * will assert LQIPHASE_LQ. Recovery is then just as |
| * listed above for the read streaming with P0 asserted. |
| * Busfree detection is enabled. |
| */ |
| if (silent == FALSE) |
| printk("LQICRC_NLQ\n"); |
| if (scb == NULL) { |
| printk("%s: No SCB valid for LQICRC_NLQ. " |
| "Resetting bus\n", ahd_name(ahd)); |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/TRUE); |
| return; |
| } |
| } else if ((lqistat1 & LQIBADLQI) != 0) { |
| printk("Need to handle BADLQI!\n"); |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/TRUE); |
| return; |
| } else if ((perrdiag & (PARITYERR|PREVPHASE)) == PARITYERR) { |
| if ((curphase & ~P_DATAIN_DT) != 0) { |
| /* Ack the byte. So we can continue. */ |
| if (silent == FALSE) |
| printk("Acking %s to clear perror\n", |
| ahd_lookup_phase_entry(curphase)->phasemsg); |
| ahd_inb(ahd, SCSIDAT); |
| } |
| |
| if (curphase == P_MESGIN) |
| msg_out = MSG_PARITY_ERROR; |
| } |
| |
| /* |
| * We've set the hardware to assert ATN if we |
| * get a parity error on "in" phases, so all we |
| * need to do is stuff the message buffer with |
| * the appropriate message. "In" phases have set |
| * mesg_out to something other than MSG_NOP. |
| */ |
| ahd->send_msg_perror = msg_out; |
| if (scb != NULL && msg_out == MSG_INITIATOR_DET_ERR) |
| scb->flags |= SCB_TRANSMISSION_ERROR; |
| ahd_outb(ahd, MSG_OUT, HOST_MSG); |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| ahd_unpause(ahd); |
| } |
| |
| static void |
| ahd_handle_lqiphase_error(struct ahd_softc *ahd, u_int lqistat1) |
| { |
| /* |
| * Clear the sources of the interrupts. |
| */ |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| ahd_outb(ahd, CLRLQIINT1, lqistat1); |
| |
| /* |
| * If the "illegal" phase changes were in response |
| * to our ATN to flag a CRC error, AND we ended up |
| * on packet boundaries, clear the error, restart the |
| * LQI manager as appropriate, and go on our merry |
| * way toward sending the message. Otherwise, reset |
| * the bus to clear the error. |
| */ |
| ahd_set_active_fifo(ahd); |
| if ((ahd_inb(ahd, SCSISIGO) & ATNO) != 0 |
| && (ahd_inb(ahd, MDFFSTAT) & DLZERO) != 0) { |
| if ((lqistat1 & LQIPHASE_LQ) != 0) { |
| printk("LQIRETRY for LQIPHASE_LQ\n"); |
| ahd_outb(ahd, LQCTL2, LQIRETRY); |
| } else if ((lqistat1 & LQIPHASE_NLQ) != 0) { |
| printk("LQIRETRY for LQIPHASE_NLQ\n"); |
| ahd_outb(ahd, LQCTL2, LQIRETRY); |
| } else |
| panic("ahd_handle_lqiphase_error: No phase errors\n"); |
| ahd_dump_card_state(ahd); |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| ahd_unpause(ahd); |
| } else { |
| printk("Resetting Channel for LQI Phase error\n"); |
| ahd_dump_card_state(ahd); |
| ahd_reset_channel(ahd, 'A', /*Initiate Reset*/TRUE); |
| } |
| } |
| |
| /* |
| * Packetized unexpected or expected busfree. |
| * Entered in mode based on busfreetime. |
| */ |
| static int |
| ahd_handle_pkt_busfree(struct ahd_softc *ahd, u_int busfreetime) |
| { |
| u_int lqostat1; |
| |
| AHD_ASSERT_MODES(ahd, ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK), |
| ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK)); |
| lqostat1 = ahd_inb(ahd, LQOSTAT1); |
| if ((lqostat1 & LQOBUSFREE) != 0) { |
| struct scb *scb; |
| u_int scbid; |
| u_int saved_scbptr; |
| u_int waiting_h; |
| u_int waiting_t; |
| u_int next; |
| |
| /* |
| * The LQO manager detected an unexpected busfree |
| * either: |
| * |
| * 1) During an outgoing LQ. |
| * 2) After an outgoing LQ but before the first |
| * REQ of the command packet. |
| * 3) During an outgoing command packet. |
| * |
| * In all cases, CURRSCB is pointing to the |
| * SCB that encountered the failure. Clean |
| * up the queue, clear SELDO and LQOBUSFREE, |
| * and allow the sequencer to restart the select |
| * out at its lesure. |
| */ |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| scbid = ahd_inw(ahd, CURRSCB); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) |
| panic("SCB not valid during LQOBUSFREE"); |
| /* |
| * Clear the status. |
| */ |
| ahd_outb(ahd, CLRLQOINT1, CLRLQOBUSFREE); |
| if ((ahd->bugs & AHD_CLRLQO_AUTOCLR_BUG) != 0) |
| ahd_outb(ahd, CLRLQOINT1, 0); |
| ahd_outb(ahd, SCSISEQ0, ahd_inb(ahd, SCSISEQ0) & ~ENSELO); |
| ahd_flush_device_writes(ahd); |
| ahd_outb(ahd, CLRSINT0, CLRSELDO); |
| |
| /* |
| * Return the LQO manager to its idle loop. It will |
| * not do this automatically if the busfree occurs |
| * after the first REQ of either the LQ or command |
| * packet or between the LQ and command packet. |
| */ |
| ahd_outb(ahd, LQCTL2, ahd_inb(ahd, LQCTL2) | LQOTOIDLE); |
| |
| /* |
| * Update the waiting for selection queue so |
| * we restart on the correct SCB. |
| */ |
| waiting_h = ahd_inw(ahd, WAITING_TID_HEAD); |
| saved_scbptr = ahd_get_scbptr(ahd); |
| if (waiting_h != scbid) { |
| |
| ahd_outw(ahd, WAITING_TID_HEAD, scbid); |
| waiting_t = ahd_inw(ahd, WAITING_TID_TAIL); |
| if (waiting_t == waiting_h) { |
| ahd_outw(ahd, WAITING_TID_TAIL, scbid); |
| next = SCB_LIST_NULL; |
| } else { |
| ahd_set_scbptr(ahd, waiting_h); |
| next = ahd_inw_scbram(ahd, SCB_NEXT2); |
| } |
| ahd_set_scbptr(ahd, scbid); |
| ahd_outw(ahd, SCB_NEXT2, next); |
| } |
| ahd_set_scbptr(ahd, saved_scbptr); |
| if (scb->crc_retry_count < AHD_MAX_LQ_CRC_ERRORS) { |
| if (SCB_IS_SILENT(scb) == FALSE) { |
| ahd_print_path(ahd, scb); |
| printk("Probable outgoing LQ CRC error. " |
| "Retrying command\n"); |
| } |
| scb->crc_retry_count++; |
| } else { |
| ahd_set_transaction_status(scb, CAM_UNCOR_PARITY); |
| ahd_freeze_scb(scb); |
| ahd_freeze_devq(ahd, scb); |
| } |
| /* Return unpausing the sequencer. */ |
| return (0); |
| } else if ((ahd_inb(ahd, PERRDIAG) & PARITYERR) != 0) { |
| /* |
| * Ignore what are really parity errors that |
| * occur on the last REQ of a free running |
| * clock prior to going busfree. Some drives |
| * do not properly active negate just before |
| * going busfree resulting in a parity glitch. |
| */ |
| ahd_outb(ahd, CLRSINT1, CLRSCSIPERR|CLRBUSFREE); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MASKED_ERRORS) != 0) |
| printk("%s: Parity on last REQ detected " |
| "during busfree phase.\n", |
| ahd_name(ahd)); |
| #endif |
| /* Return unpausing the sequencer. */ |
| return (0); |
| } |
| if (ahd->src_mode != AHD_MODE_SCSI) { |
| u_int scbid; |
| struct scb *scb; |
| |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| ahd_print_path(ahd, scb); |
| printk("Unexpected PKT busfree condition\n"); |
| ahd_dump_card_state(ahd); |
| ahd_abort_scbs(ahd, SCB_GET_TARGET(ahd, scb), 'A', |
| SCB_GET_LUN(scb), SCB_GET_TAG(scb), |
| ROLE_INITIATOR, CAM_UNEXP_BUSFREE); |
| |
| /* Return restarting the sequencer. */ |
| return (1); |
| } |
| printk("%s: Unexpected PKT busfree condition\n", ahd_name(ahd)); |
| ahd_dump_card_state(ahd); |
| /* Restart the sequencer. */ |
| return (1); |
| } |
| |
| /* |
| * Non-packetized unexpected or expected busfree. |
| */ |
| static int |
| ahd_handle_nonpkt_busfree(struct ahd_softc *ahd) |
| { |
| struct ahd_devinfo devinfo; |
| struct scb *scb; |
| u_int lastphase; |
| u_int saved_scsiid; |
| u_int saved_lun; |
| u_int target; |
| u_int initiator_role_id; |
| u_int scbid; |
| u_int ppr_busfree; |
| int printerror; |
| |
| /* |
| * Look at what phase we were last in. If its message out, |
| * chances are pretty good that the busfree was in response |
| * to one of our abort requests. |
| */ |
| lastphase = ahd_inb(ahd, LASTPHASE); |
| saved_scsiid = ahd_inb(ahd, SAVED_SCSIID); |
| saved_lun = ahd_inb(ahd, SAVED_LUN); |
| target = SCSIID_TARGET(ahd, saved_scsiid); |
| initiator_role_id = SCSIID_OUR_ID(saved_scsiid); |
| ahd_compile_devinfo(&devinfo, initiator_role_id, |
| target, saved_lun, 'A', ROLE_INITIATOR); |
| printerror = 1; |
| |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb != NULL |
| && (ahd_inb(ahd, SEQ_FLAGS) & NOT_IDENTIFIED) != 0) |
| scb = NULL; |
| |
| ppr_busfree = (ahd->msg_flags & MSG_FLAG_EXPECT_PPR_BUSFREE) != 0; |
| if (lastphase == P_MESGOUT) { |
| u_int tag; |
| |
| tag = SCB_LIST_NULL; |
| if (ahd_sent_msg(ahd, AHDMSG_1B, MSG_ABORT_TAG, TRUE) |
| || ahd_sent_msg(ahd, AHDMSG_1B, MSG_ABORT, TRUE)) { |
| int found; |
| int sent_msg; |
| |
| if (scb == NULL) { |
| ahd_print_devinfo(ahd, &devinfo); |
| printk("Abort for unidentified " |
| "connection completed.\n"); |
| /* restart the sequencer. */ |
| return (1); |
| } |
| sent_msg = ahd->msgout_buf[ahd->msgout_index - 1]; |
| ahd_print_path(ahd, scb); |
| printk("SCB %d - Abort%s Completed.\n", |
| SCB_GET_TAG(scb), |
| sent_msg == MSG_ABORT_TAG ? "" : " Tag"); |
| |
| if (sent_msg == MSG_ABORT_TAG) |
| tag = SCB_GET_TAG(scb); |
| |
| if ((scb->flags & SCB_EXTERNAL_RESET) != 0) { |
| /* |
| * This abort is in response to an |
| * unexpected switch to command phase |
| * for a packetized connection. Since |
| * the identify message was never sent, |
| * "saved lun" is 0. We really want to |
| * abort only the SCB that encountered |
| * this error, which could have a different |
| * lun. The SCB will be retried so the OS |
| * will see the UA after renegotiating to |
| * packetized. |
| */ |
| tag = SCB_GET_TAG(scb); |
| saved_lun = scb->hscb->lun; |
| } |
| found = ahd_abort_scbs(ahd, target, 'A', saved_lun, |
| tag, ROLE_INITIATOR, |
| CAM_REQ_ABORTED); |
| printk("found == 0x%x\n", found); |
| printerror = 0; |
| } else if (ahd_sent_msg(ahd, AHDMSG_1B, |
| MSG_BUS_DEV_RESET, TRUE)) { |
| #ifdef __FreeBSD__ |
| /* |
| * Don't mark the user's request for this BDR |
| * as completing with CAM_BDR_SENT. CAM3 |
| * specifies CAM_REQ_CMP. |
| */ |
| if (scb != NULL |
| && scb->io_ctx->ccb_h.func_code== XPT_RESET_DEV |
| && ahd_match_scb(ahd, scb, target, 'A', |
| CAM_LUN_WILDCARD, SCB_LIST_NULL, |
| ROLE_INITIATOR)) |
| ahd_set_transaction_status(scb, CAM_REQ_CMP); |
| #endif |
| ahd_handle_devreset(ahd, &devinfo, CAM_LUN_WILDCARD, |
| CAM_BDR_SENT, "Bus Device Reset", |
| /*verbose_level*/0); |
| printerror = 0; |
| } else if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_PPR, FALSE) |
| && ppr_busfree == 0) { |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_tmode_tstate *tstate; |
| |
| /* |
| * PPR Rejected. |
| * |
| * If the previous negotiation was packetized, |
| * this could be because the device has been |
| * reset without our knowledge. Force our |
| * current negotiation to async and retry the |
| * negotiation. Otherwise retry the command |
| * with non-ppr negotiation. |
| */ |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk("PPR negotiation rejected busfree.\n"); |
| #endif |
| tinfo = ahd_fetch_transinfo(ahd, devinfo.channel, |
| devinfo.our_scsiid, |
| devinfo.target, &tstate); |
| if ((tinfo->curr.ppr_options & MSG_EXT_PPR_IU_REQ)!=0) { |
| ahd_set_width(ahd, &devinfo, |
| MSG_EXT_WDTR_BUS_8_BIT, |
| AHD_TRANS_CUR, |
| /*paused*/TRUE); |
| ahd_set_syncrate(ahd, &devinfo, |
| /*period*/0, /*offset*/0, |
| /*ppr_options*/0, |
| AHD_TRANS_CUR, |
| /*paused*/TRUE); |
| /* |
| * The expect PPR busfree handler below |
| * will effect the retry and necessary |
| * abort. |
| */ |
| } else { |
| tinfo->curr.transport_version = 2; |
| tinfo->goal.transport_version = 2; |
| tinfo->goal.ppr_options = 0; |
| if (scb != NULL) { |
| /* |
| * Remove any SCBs in the waiting |
| * for selection queue that may |
| * also be for this target so that |
| * command ordering is preserved. |
| */ |
| ahd_freeze_devq(ahd, scb); |
| ahd_qinfifo_requeue_tail(ahd, scb); |
| } |
| printerror = 0; |
| } |
| } else if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_WDTR, FALSE) |
| && ppr_busfree == 0) { |
| /* |
| * Negotiation Rejected. Go-narrow and |
| * retry command. |
| */ |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk("WDTR negotiation rejected busfree.\n"); |
| #endif |
| ahd_set_width(ahd, &devinfo, |
| MSG_EXT_WDTR_BUS_8_BIT, |
| AHD_TRANS_CUR|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| if (scb != NULL) { |
| /* |
| * Remove any SCBs in the waiting for |
| * selection queue that may also be for |
| * this target so that command ordering |
| * is preserved. |
| */ |
| ahd_freeze_devq(ahd, scb); |
| ahd_qinfifo_requeue_tail(ahd, scb); |
| } |
| printerror = 0; |
| } else if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_SDTR, FALSE) |
| && ppr_busfree == 0) { |
| /* |
| * Negotiation Rejected. Go-async and |
| * retry command. |
| */ |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk("SDTR negotiation rejected busfree.\n"); |
| #endif |
| ahd_set_syncrate(ahd, &devinfo, |
| /*period*/0, /*offset*/0, |
| /*ppr_options*/0, |
| AHD_TRANS_CUR|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| if (scb != NULL) { |
| /* |
| * Remove any SCBs in the waiting for |
| * selection queue that may also be for |
| * this target so that command ordering |
| * is preserved. |
| */ |
| ahd_freeze_devq(ahd, scb); |
| ahd_qinfifo_requeue_tail(ahd, scb); |
| } |
| printerror = 0; |
| } else if ((ahd->msg_flags & MSG_FLAG_EXPECT_IDE_BUSFREE) != 0 |
| && ahd_sent_msg(ahd, AHDMSG_1B, |
| MSG_INITIATOR_DET_ERR, TRUE)) { |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk("Expected IDE Busfree\n"); |
| #endif |
| printerror = 0; |
| } else if ((ahd->msg_flags & MSG_FLAG_EXPECT_QASREJ_BUSFREE) |
| && ahd_sent_msg(ahd, AHDMSG_1B, |
| MSG_MESSAGE_REJECT, TRUE)) { |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk("Expected QAS Reject Busfree\n"); |
| #endif |
| printerror = 0; |
| } |
| } |
| |
| /* |
| * The busfree required flag is honored at the end of |
| * the message phases. We check it last in case we |
| * had to send some other message that caused a busfree. |
| */ |
| if (scb != NULL && printerror != 0 |
| && (lastphase == P_MESGIN || lastphase == P_MESGOUT) |
| && ((ahd->msg_flags & MSG_FLAG_EXPECT_PPR_BUSFREE) != 0)) { |
| |
| ahd_freeze_devq(ahd, scb); |
| ahd_set_transaction_status(scb, CAM_REQUEUE_REQ); |
| ahd_freeze_scb(scb); |
| if ((ahd->msg_flags & MSG_FLAG_IU_REQ_CHANGED) != 0) { |
| ahd_abort_scbs(ahd, SCB_GET_TARGET(ahd, scb), |
| SCB_GET_CHANNEL(ahd, scb), |
| SCB_GET_LUN(scb), SCB_LIST_NULL, |
| ROLE_INITIATOR, CAM_REQ_ABORTED); |
| } else { |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk("PPR Negotiation Busfree.\n"); |
| #endif |
| ahd_done(ahd, scb); |
| } |
| printerror = 0; |
| } |
| if (printerror != 0) { |
| int aborted; |
| |
| aborted = 0; |
| if (scb != NULL) { |
| u_int tag; |
| |
| if ((scb->hscb->control & TAG_ENB) != 0) |
| tag = SCB_GET_TAG(scb); |
| else |
| tag = SCB_LIST_NULL; |
| ahd_print_path(ahd, scb); |
| aborted = ahd_abort_scbs(ahd, target, 'A', |
| SCB_GET_LUN(scb), tag, |
| ROLE_INITIATOR, |
| CAM_UNEXP_BUSFREE); |
| } else { |
| /* |
| * We had not fully identified this connection, |
| * so we cannot abort anything. |
| */ |
| printk("%s: ", ahd_name(ahd)); |
| } |
| printk("Unexpected busfree %s, %d SCBs aborted, " |
| "PRGMCNT == 0x%x\n", |
| ahd_lookup_phase_entry(lastphase)->phasemsg, |
| aborted, |
| ahd_inw(ahd, PRGMCNT)); |
| ahd_dump_card_state(ahd); |
| if (lastphase != P_BUSFREE) |
| ahd_force_renegotiation(ahd, &devinfo); |
| } |
| /* Always restart the sequencer. */ |
| return (1); |
| } |
| |
| static void |
| ahd_handle_proto_violation(struct ahd_softc *ahd) |
| { |
| struct ahd_devinfo devinfo; |
| struct scb *scb; |
| u_int scbid; |
| u_int seq_flags; |
| u_int curphase; |
| u_int lastphase; |
| int found; |
| |
| ahd_fetch_devinfo(ahd, &devinfo); |
| scbid = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scbid); |
| seq_flags = ahd_inb(ahd, SEQ_FLAGS); |
| curphase = ahd_inb(ahd, SCSISIGI) & PHASE_MASK; |
| lastphase = ahd_inb(ahd, LASTPHASE); |
| if ((seq_flags & NOT_IDENTIFIED) != 0) { |
| |
| /* |
| * The reconnecting target either did not send an |
| * identify message, or did, but we didn't find an SCB |
| * to match. |
| */ |
| ahd_print_devinfo(ahd, &devinfo); |
| printk("Target did not send an IDENTIFY message. " |
| "LASTPHASE = 0x%x.\n", lastphase); |
| scb = NULL; |
| } else if (scb == NULL) { |
| /* |
| * We don't seem to have an SCB active for this |
| * transaction. Print an error and reset the bus. |
| */ |
| ahd_print_devinfo(ahd, &devinfo); |
| printk("No SCB found during protocol violation\n"); |
| goto proto_violation_reset; |
| } else { |
| ahd_set_transaction_status(scb, CAM_SEQUENCE_FAIL); |
| if ((seq_flags & NO_CDB_SENT) != 0) { |
| ahd_print_path(ahd, scb); |
| printk("No or incomplete CDB sent to device.\n"); |
| } else if ((ahd_inb_scbram(ahd, SCB_CONTROL) |
| & STATUS_RCVD) == 0) { |
| /* |
| * The target never bothered to provide status to |
| * us prior to completing the command. Since we don't |
| * know the disposition of this command, we must attempt |
| * to abort it. Assert ATN and prepare to send an abort |
| * message. |
| */ |
| ahd_print_path(ahd, scb); |
| printk("Completed command without status.\n"); |
| } else { |
| ahd_print_path(ahd, scb); |
| printk("Unknown protocol violation.\n"); |
| ahd_dump_card_state(ahd); |
| } |
| } |
| if ((lastphase & ~P_DATAIN_DT) == 0 |
| || lastphase == P_COMMAND) { |
| proto_violation_reset: |
| /* |
| * Target either went directly to data |
| * phase or didn't respond to our ATN. |
| * The only safe thing to do is to blow |
| * it away with a bus reset. |
| */ |
| found = ahd_reset_channel(ahd, 'A', TRUE); |
| printk("%s: Issued Channel %c Bus Reset. " |
| "%d SCBs aborted\n", ahd_name(ahd), 'A', found); |
| } else { |
| /* |
| * Leave the selection hardware off in case |
| * this abort attempt will affect yet to |
| * be sent commands. |
| */ |
| ahd_outb(ahd, SCSISEQ0, |
| ahd_inb(ahd, SCSISEQ0) & ~ENSELO); |
| ahd_assert_atn(ahd); |
| ahd_outb(ahd, MSG_OUT, HOST_MSG); |
| if (scb == NULL) { |
| ahd_print_devinfo(ahd, &devinfo); |
| ahd->msgout_buf[0] = MSG_ABORT_TASK; |
| ahd->msgout_len = 1; |
| ahd->msgout_index = 0; |
| ahd->msg_type = MSG_TYPE_INITIATOR_MSGOUT; |
| } else { |
| ahd_print_path(ahd, scb); |
| scb->flags |= SCB_ABORT; |
| } |
| printk("Protocol violation %s. Attempting to abort.\n", |
| ahd_lookup_phase_entry(curphase)->phasemsg); |
| } |
| } |
| |
| /* |
| * Force renegotiation to occur the next time we initiate |
| * a command to the current device. |
| */ |
| static void |
| ahd_force_renegotiation(struct ahd_softc *ahd, struct ahd_devinfo *devinfo) |
| { |
| struct ahd_initiator_tinfo *targ_info; |
| struct ahd_tmode_tstate *tstate; |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) { |
| ahd_print_devinfo(ahd, devinfo); |
| printk("Forcing renegotiation\n"); |
| } |
| #endif |
| targ_info = ahd_fetch_transinfo(ahd, |
| devinfo->channel, |
| devinfo->our_scsiid, |
| devinfo->target, |
| &tstate); |
| ahd_update_neg_request(ahd, devinfo, tstate, |
| targ_info, AHD_NEG_IF_NON_ASYNC); |
| } |
| |
| #define AHD_MAX_STEPS 2000 |
| static void |
| ahd_clear_critical_section(struct ahd_softc *ahd) |
| { |
| ahd_mode_state saved_modes; |
| int stepping; |
| int steps; |
| int first_instr; |
| u_int simode0; |
| u_int simode1; |
| u_int simode3; |
| u_int lqimode0; |
| u_int lqimode1; |
| u_int lqomode0; |
| u_int lqomode1; |
| |
| if (ahd->num_critical_sections == 0) |
| return; |
| |
| stepping = FALSE; |
| steps = 0; |
| first_instr = 0; |
| simode0 = 0; |
| simode1 = 0; |
| simode3 = 0; |
| lqimode0 = 0; |
| lqimode1 = 0; |
| lqomode0 = 0; |
| lqomode1 = 0; |
| saved_modes = ahd_save_modes(ahd); |
| for (;;) { |
| struct cs *cs; |
| u_int seqaddr; |
| u_int i; |
| |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| seqaddr = ahd_inw(ahd, CURADDR); |
| |
| cs = ahd->critical_sections; |
| for (i = 0; i < ahd->num_critical_sections; i++, cs++) { |
| |
| if (cs->begin < seqaddr && cs->end >= seqaddr) |
| break; |
| } |
| |
| if (i == ahd->num_critical_sections) |
| break; |
| |
| if (steps > AHD_MAX_STEPS) { |
| printk("%s: Infinite loop in critical section\n" |
| "%s: First Instruction 0x%x now 0x%x\n", |
| ahd_name(ahd), ahd_name(ahd), first_instr, |
| seqaddr); |
| ahd_dump_card_state(ahd); |
| panic("critical section loop"); |
| } |
| |
| steps++; |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MISC) != 0) |
| printk("%s: Single stepping at 0x%x\n", ahd_name(ahd), |
| seqaddr); |
| #endif |
| if (stepping == FALSE) { |
| |
| first_instr = seqaddr; |
| ahd_set_modes(ahd, AHD_MODE_CFG, AHD_MODE_CFG); |
| simode0 = ahd_inb(ahd, SIMODE0); |
| simode3 = ahd_inb(ahd, SIMODE3); |
| lqimode0 = ahd_inb(ahd, LQIMODE0); |
| lqimode1 = ahd_inb(ahd, LQIMODE1); |
| lqomode0 = ahd_inb(ahd, LQOMODE0); |
| lqomode1 = ahd_inb(ahd, LQOMODE1); |
| ahd_outb(ahd, SIMODE0, 0); |
| ahd_outb(ahd, SIMODE3, 0); |
| ahd_outb(ahd, LQIMODE0, 0); |
| ahd_outb(ahd, LQIMODE1, 0); |
| ahd_outb(ahd, LQOMODE0, 0); |
| ahd_outb(ahd, LQOMODE1, 0); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| simode1 = ahd_inb(ahd, SIMODE1); |
| /* |
| * We don't clear ENBUSFREE. Unfortunately |
| * we cannot re-enable busfree detection within |
| * the current connection, so we must leave it |
| * on while single stepping. |
| */ |
| ahd_outb(ahd, SIMODE1, simode1 & ENBUSFREE); |
| ahd_outb(ahd, SEQCTL0, ahd_inb(ahd, SEQCTL0) | STEP); |
| stepping = TRUE; |
| } |
| ahd_outb(ahd, CLRSINT1, CLRBUSFREE); |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| ahd_set_modes(ahd, ahd->saved_src_mode, ahd->saved_dst_mode); |
| ahd_outb(ahd, HCNTRL, ahd->unpause); |
| while (!ahd_is_paused(ahd)) |
| ahd_delay(200); |
| ahd_update_modes(ahd); |
| } |
| if (stepping) { |
| ahd_set_modes(ahd, AHD_MODE_CFG, AHD_MODE_CFG); |
| ahd_outb(ahd, SIMODE0, simode0); |
| ahd_outb(ahd, SIMODE3, simode3); |
| ahd_outb(ahd, LQIMODE0, lqimode0); |
| ahd_outb(ahd, LQIMODE1, lqimode1); |
| ahd_outb(ahd, LQOMODE0, lqomode0); |
| ahd_outb(ahd, LQOMODE1, lqomode1); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| ahd_outb(ahd, SEQCTL0, ahd_inb(ahd, SEQCTL0) & ~STEP); |
| ahd_outb(ahd, SIMODE1, simode1); |
| /* |
| * SCSIINT seems to glitch occasionally when |
| * the interrupt masks are restored. Clear SCSIINT |
| * one more time so that only persistent errors |
| * are seen as a real interrupt. |
| */ |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| } |
| ahd_restore_modes(ahd, saved_modes); |
| } |
| |
| /* |
| * Clear any pending interrupt status. |
| */ |
| static void |
| ahd_clear_intstat(struct ahd_softc *ahd) |
| { |
| AHD_ASSERT_MODES(ahd, ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK), |
| ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK)); |
| /* Clear any interrupt conditions this may have caused */ |
| ahd_outb(ahd, CLRLQIINT0, CLRLQIATNQAS|CLRLQICRCT1|CLRLQICRCT2 |
| |CLRLQIBADLQT|CLRLQIATNLQ|CLRLQIATNCMD); |
| ahd_outb(ahd, CLRLQIINT1, CLRLQIPHASE_LQ|CLRLQIPHASE_NLQ|CLRLIQABORT |
| |CLRLQICRCI_LQ|CLRLQICRCI_NLQ|CLRLQIBADLQI |
| |CLRLQIOVERI_LQ|CLRLQIOVERI_NLQ|CLRNONPACKREQ); |
| ahd_outb(ahd, CLRLQOINT0, CLRLQOTARGSCBPERR|CLRLQOSTOPT2|CLRLQOATNLQ |
| |CLRLQOATNPKT|CLRLQOTCRC); |
| ahd_outb(ahd, CLRLQOINT1, CLRLQOINITSCBPERR|CLRLQOSTOPI2|CLRLQOBADQAS |
| |CLRLQOBUSFREE|CLRLQOPHACHGINPKT); |
| if ((ahd->bugs & AHD_CLRLQO_AUTOCLR_BUG) != 0) { |
| ahd_outb(ahd, CLRLQOINT0, 0); |
| ahd_outb(ahd, CLRLQOINT1, 0); |
| } |
| ahd_outb(ahd, CLRSINT3, CLRNTRAMPERR|CLROSRAMPERR); |
| ahd_outb(ahd, CLRSINT1, CLRSELTIMEO|CLRATNO|CLRSCSIRSTI |
| |CLRBUSFREE|CLRSCSIPERR|CLRREQINIT); |
| ahd_outb(ahd, CLRSINT0, CLRSELDO|CLRSELDI|CLRSELINGO |
| |CLRIOERR|CLROVERRUN); |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| } |
| |
| /**************************** Debugging Routines ******************************/ |
| #ifdef AHD_DEBUG |
| uint32_t ahd_debug = AHD_DEBUG_OPTS; |
| #endif |
| |
| #if 0 |
| void |
| ahd_print_scb(struct scb *scb) |
| { |
| struct hardware_scb *hscb; |
| int i; |
| |
| hscb = scb->hscb; |
| printk("scb:%p control:0x%x scsiid:0x%x lun:%d cdb_len:%d\n", |
| (void *)scb, |
| hscb->control, |
| hscb->scsiid, |
| hscb->lun, |
| hscb->cdb_len); |
| printk("Shared Data: "); |
| for (i = 0; i < sizeof(hscb->shared_data.idata.cdb); i++) |
| printk("%#02x", hscb->shared_data.idata.cdb[i]); |
| printk(" dataptr:%#x%x datacnt:%#x sgptr:%#x tag:%#x\n", |
| (uint32_t)((ahd_le64toh(hscb->dataptr) >> 32) & 0xFFFFFFFF), |
| (uint32_t)(ahd_le64toh(hscb->dataptr) & 0xFFFFFFFF), |
| ahd_le32toh(hscb->datacnt), |
| ahd_le32toh(hscb->sgptr), |
| SCB_GET_TAG(scb)); |
| ahd_dump_sglist(scb); |
| } |
| #endif /* 0 */ |
| |
| /************************* Transfer Negotiation *******************************/ |
| /* |
| * Allocate per target mode instance (ID we respond to as a target) |
| * transfer negotiation data structures. |
| */ |
| static struct ahd_tmode_tstate * |
| ahd_alloc_tstate(struct ahd_softc *ahd, u_int scsi_id, char channel) |
| { |
| struct ahd_tmode_tstate *master_tstate; |
| struct ahd_tmode_tstate *tstate; |
| int i; |
| |
| master_tstate = ahd->enabled_targets[ahd->our_id]; |
| if (ahd->enabled_targets[scsi_id] != NULL |
| && ahd->enabled_targets[scsi_id] != master_tstate) |
| panic("%s: ahd_alloc_tstate - Target already allocated", |
| ahd_name(ahd)); |
| tstate = kmalloc(sizeof(*tstate), GFP_ATOMIC); |
| if (tstate == NULL) |
| return (NULL); |
| |
| /* |
| * If we have allocated a master tstate, copy user settings from |
| * the master tstate (taken from SRAM or the EEPROM) for this |
| * channel, but reset our current and goal settings to async/narrow |
| * until an initiator talks to us. |
| */ |
| if (master_tstate != NULL) { |
| memcpy(tstate, master_tstate, sizeof(*tstate)); |
| memset(tstate->enabled_luns, 0, sizeof(tstate->enabled_luns)); |
| for (i = 0; i < 16; i++) { |
| memset(&tstate->transinfo[i].curr, 0, |
| sizeof(tstate->transinfo[i].curr)); |
| memset(&tstate->transinfo[i].goal, 0, |
| sizeof(tstate->transinfo[i].goal)); |
| } |
| } else |
| memset(tstate, 0, sizeof(*tstate)); |
| ahd->enabled_targets[scsi_id] = tstate; |
| return (tstate); |
| } |
| |
| #ifdef AHD_TARGET_MODE |
| /* |
| * Free per target mode instance (ID we respond to as a target) |
| * transfer negotiation data structures. |
| */ |
| static void |
| ahd_free_tstate(struct ahd_softc *ahd, u_int scsi_id, char channel, int force) |
| { |
| struct ahd_tmode_tstate *tstate; |
| |
| /* |
| * Don't clean up our "master" tstate. |
| * It has our default user settings. |
| */ |
| if (scsi_id == ahd->our_id |
| && force == FALSE) |
| return; |
| |
| tstate = ahd->enabled_targets[scsi_id]; |
| if (tstate != NULL) |
| kfree(tstate); |
| ahd->enabled_targets[scsi_id] = NULL; |
| } |
| #endif |
| |
| /* |
| * Called when we have an active connection to a target on the bus, |
| * this function finds the nearest period to the input period limited |
| * by the capabilities of the bus connectivity of and sync settings for |
| * the target. |
| */ |
| static void |
| ahd_devlimited_syncrate(struct ahd_softc *ahd, |
| struct ahd_initiator_tinfo *tinfo, |
| u_int *period, u_int *ppr_options, role_t role) |
| { |
| struct ahd_transinfo *transinfo; |
| u_int maxsync; |
| |
| if ((ahd_inb(ahd, SBLKCTL) & ENAB40) != 0 |
| && (ahd_inb(ahd, SSTAT2) & EXP_ACTIVE) == 0) { |
| maxsync = AHD_SYNCRATE_PACED; |
| } else { |
| maxsync = AHD_SYNCRATE_ULTRA; |
| /* Can't do DT related options on an SE bus */ |
| *ppr_options &= MSG_EXT_PPR_QAS_REQ; |
| } |
| /* |
| * Never allow a value higher than our current goal |
| * period otherwise we may allow a target initiated |
| * negotiation to go above the limit as set by the |
| * user. In the case of an initiator initiated |
| * sync negotiation, we limit based on the user |
| * setting. This allows the system to still accept |
| * incoming negotiations even if target initiated |
| * negotiation is not performed. |
| */ |
| if (role == ROLE_TARGET) |
| transinfo = &tinfo->user; |
| else |
| transinfo = &tinfo->goal; |
| *ppr_options &= (transinfo->ppr_options|MSG_EXT_PPR_PCOMP_EN); |
| if (transinfo->width == MSG_EXT_WDTR_BUS_8_BIT) { |
| maxsync = max(maxsync, (u_int)AHD_SYNCRATE_ULTRA2); |
| *ppr_options &= ~MSG_EXT_PPR_DT_REQ; |
| } |
| if (transinfo->period == 0) { |
| *period = 0; |
| *ppr_options = 0; |
| } else { |
| *period = max(*period, (u_int)transinfo->period); |
| ahd_find_syncrate(ahd, period, ppr_options, maxsync); |
| } |
| } |
| |
| /* |
| * Look up the valid period to SCSIRATE conversion in our table. |
| * Return the period and offset that should be sent to the target |
| * if this was the beginning of an SDTR. |
| */ |
| void |
| ahd_find_syncrate(struct ahd_softc *ahd, u_int *period, |
| u_int *ppr_options, u_int maxsync) |
| { |
| if (*period < maxsync) |
| *period = maxsync; |
| |
| if ((*ppr_options & MSG_EXT_PPR_DT_REQ) != 0 |
| && *period > AHD_SYNCRATE_MIN_DT) |
| *ppr_options &= ~MSG_EXT_PPR_DT_REQ; |
| |
| if (*period > AHD_SYNCRATE_MIN) |
| *period = 0; |
| |
| /* Honor PPR option conformance rules. */ |
| if (*period > AHD_SYNCRATE_PACED) |
| *ppr_options &= ~MSG_EXT_PPR_RTI; |
| |
| if ((*ppr_options & MSG_EXT_PPR_IU_REQ) == 0) |
| *ppr_options &= (MSG_EXT_PPR_DT_REQ|MSG_EXT_PPR_QAS_REQ); |
| |
| if ((*ppr_options & MSG_EXT_PPR_DT_REQ) == 0) |
| *ppr_options &= MSG_EXT_PPR_QAS_REQ; |
| |
| /* Skip all PACED only entries if IU is not available */ |
| if ((*ppr_options & MSG_EXT_PPR_IU_REQ) == 0 |
| && *period < AHD_SYNCRATE_DT) |
| *period = AHD_SYNCRATE_DT; |
| |
| /* Skip all DT only entries if DT is not available */ |
| if ((*ppr_options & MSG_EXT_PPR_DT_REQ) == 0 |
| && *period < AHD_SYNCRATE_ULTRA2) |
| *period = AHD_SYNCRATE_ULTRA2; |
| } |
| |
| /* |
| * Truncate the given synchronous offset to a value the |
| * current adapter type and syncrate are capable of. |
| */ |
| static void |
| ahd_validate_offset(struct ahd_softc *ahd, |
| struct ahd_initiator_tinfo *tinfo, |
| u_int period, u_int *offset, int wide, |
| role_t role) |
| { |
| u_int maxoffset; |
| |
| /* Limit offset to what we can do */ |
| if (period == 0) |
| maxoffset = 0; |
| else if (period <= AHD_SYNCRATE_PACED) { |
| if ((ahd->bugs & AHD_PACED_NEGTABLE_BUG) != 0) |
| maxoffset = MAX_OFFSET_PACED_BUG; |
| else |
| maxoffset = MAX_OFFSET_PACED; |
| } else |
| maxoffset = MAX_OFFSET_NON_PACED; |
| *offset = min(*offset, maxoffset); |
| if (tinfo != NULL) { |
| if (role == ROLE_TARGET) |
| *offset = min(*offset, (u_int)tinfo->user.offset); |
| else |
| *offset = min(*offset, (u_int)tinfo->goal.offset); |
| } |
| } |
| |
| /* |
| * Truncate the given transfer width parameter to a value the |
| * current adapter type is capable of. |
| */ |
| static void |
| ahd_validate_width(struct ahd_softc *ahd, struct ahd_initiator_tinfo *tinfo, |
| u_int *bus_width, role_t role) |
| { |
| switch (*bus_width) { |
| default: |
| if (ahd->features & AHD_WIDE) { |
| /* Respond Wide */ |
| *bus_width = MSG_EXT_WDTR_BUS_16_BIT; |
| break; |
| } |
| /* FALLTHROUGH */ |
| case MSG_EXT_WDTR_BUS_8_BIT: |
| *bus_width = MSG_EXT_WDTR_BUS_8_BIT; |
| break; |
| } |
| if (tinfo != NULL) { |
| if (role == ROLE_TARGET) |
| *bus_width = min((u_int)tinfo->user.width, *bus_width); |
| else |
| *bus_width = min((u_int)tinfo->goal.width, *bus_width); |
| } |
| } |
| |
| /* |
| * Update the bitmask of targets for which the controller should |
| * negotiate with at the next convenient opportunity. This currently |
| * means the next time we send the initial identify messages for |
| * a new transaction. |
| */ |
| int |
| ahd_update_neg_request(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| struct ahd_tmode_tstate *tstate, |
| struct ahd_initiator_tinfo *tinfo, ahd_neg_type neg_type) |
| { |
| u_int auto_negotiate_orig; |
| |
| auto_negotiate_orig = tstate->auto_negotiate; |
| if (neg_type == AHD_NEG_ALWAYS) { |
| /* |
| * Force our "current" settings to be |
| * unknown so that unless a bus reset |
| * occurs the need to renegotiate is |
| * recorded persistently. |
| */ |
| if ((ahd->features & AHD_WIDE) != 0) |
| tinfo->curr.width = AHD_WIDTH_UNKNOWN; |
| tinfo->curr.period = AHD_PERIOD_UNKNOWN; |
| tinfo->curr.offset = AHD_OFFSET_UNKNOWN; |
| } |
| if (tinfo->curr.period != tinfo->goal.period |
| || tinfo->curr.width != tinfo->goal.width |
| || tinfo->curr.offset != tinfo->goal.offset |
| || tinfo->curr.ppr_options != tinfo->goal.ppr_options |
| || (neg_type == AHD_NEG_IF_NON_ASYNC |
| && (tinfo->goal.offset != 0 |
| || tinfo->goal.width != MSG_EXT_WDTR_BUS_8_BIT |
| || tinfo->goal.ppr_options != 0))) |
| tstate->auto_negotiate |= devinfo->target_mask; |
| else |
| tstate->auto_negotiate &= ~devinfo->target_mask; |
| |
| return (auto_negotiate_orig != tstate->auto_negotiate); |
| } |
| |
| /* |
| * Update the user/goal/curr tables of synchronous negotiation |
| * parameters as well as, in the case of a current or active update, |
| * any data structures on the host controller. In the case of an |
| * active update, the specified target is currently talking to us on |
| * the bus, so the transfer parameter update must take effect |
| * immediately. |
| */ |
| void |
| ahd_set_syncrate(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| u_int period, u_int offset, u_int ppr_options, |
| u_int type, int paused) |
| { |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_tmode_tstate *tstate; |
| u_int old_period; |
| u_int old_offset; |
| u_int old_ppr; |
| int active; |
| int update_needed; |
| |
| active = (type & AHD_TRANS_ACTIVE) == AHD_TRANS_ACTIVE; |
| update_needed = 0; |
| |
| if (period == 0 || offset == 0) { |
| period = 0; |
| offset = 0; |
| } |
| |
| tinfo = ahd_fetch_transinfo(ahd, devinfo->channel, devinfo->our_scsiid, |
| devinfo->target, &tstate); |
| |
| if ((type & AHD_TRANS_USER) != 0) { |
| tinfo->user.period = period; |
| tinfo->user.offset = offset; |
| tinfo->user.ppr_options = ppr_options; |
| } |
| |
| if ((type & AHD_TRANS_GOAL) != 0) { |
| tinfo->goal.period = period; |
| tinfo->goal.offset = offset; |
| tinfo->goal.ppr_options = ppr_options; |
| } |
| |
| old_period = tinfo->curr.period; |
| old_offset = tinfo->curr.offset; |
| old_ppr = tinfo->curr.ppr_options; |
| |
| if ((type & AHD_TRANS_CUR) != 0 |
| && (old_period != period |
| || old_offset != offset |
| || old_ppr != ppr_options)) { |
| |
| update_needed++; |
| |
| tinfo->curr.period = period; |
| tinfo->curr.offset = offset; |
| tinfo->curr.ppr_options = ppr_options; |
| |
| ahd_send_async(ahd, devinfo->channel, devinfo->target, |
| CAM_LUN_WILDCARD, AC_TRANSFER_NEG); |
| if (bootverbose) { |
| if (offset != 0) { |
| int options; |
| |
| printk("%s: target %d synchronous with " |
| "period = 0x%x, offset = 0x%x", |
| ahd_name(ahd), devinfo->target, |
| period, offset); |
| options = 0; |
| if ((ppr_options & MSG_EXT_PPR_RD_STRM) != 0) { |
| printk("(RDSTRM"); |
| options++; |
| } |
| if ((ppr_options & MSG_EXT_PPR_DT_REQ) != 0) { |
| printk("%s", options ? "|DT" : "(DT"); |
| options++; |
| } |
| if ((ppr_options & MSG_EXT_PPR_IU_REQ) != 0) { |
| printk("%s", options ? "|IU" : "(IU"); |
| options++; |
| } |
| if ((ppr_options & MSG_EXT_PPR_RTI) != 0) { |
| printk("%s", options ? "|RTI" : "(RTI"); |
| options++; |
| } |
| if ((ppr_options & MSG_EXT_PPR_QAS_REQ) != 0) { |
| printk("%s", options ? "|QAS" : "(QAS"); |
| options++; |
| } |
| if (options != 0) |
| printk(")\n"); |
| else |
| printk("\n"); |
| } else { |
| printk("%s: target %d using " |
| "asynchronous transfers%s\n", |
| ahd_name(ahd), devinfo->target, |
| (ppr_options & MSG_EXT_PPR_QAS_REQ) != 0 |
| ? "(QAS)" : ""); |
| } |
| } |
| } |
| /* |
| * Always refresh the neg-table to handle the case of the |
| * sequencer setting the ENATNO bit for a MK_MESSAGE request. |
| * We will always renegotiate in that case if this is a |
| * packetized request. Also manage the busfree expected flag |
| * from this common routine so that we catch changes due to |
| * WDTR or SDTR messages. |
| */ |
| if ((type & AHD_TRANS_CUR) != 0) { |
| if (!paused) |
| ahd_pause(ahd); |
| ahd_update_neg_table(ahd, devinfo, &tinfo->curr); |
| if (!paused) |
| ahd_unpause(ahd); |
| if (ahd->msg_type != MSG_TYPE_NONE) { |
| if ((old_ppr & MSG_EXT_PPR_IU_REQ) |
| != (ppr_options & MSG_EXT_PPR_IU_REQ)) { |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) { |
| ahd_print_devinfo(ahd, devinfo); |
| printk("Expecting IU Change busfree\n"); |
| } |
| #endif |
| ahd->msg_flags |= MSG_FLAG_EXPECT_PPR_BUSFREE |
| | MSG_FLAG_IU_REQ_CHANGED; |
| } |
| if ((old_ppr & MSG_EXT_PPR_IU_REQ) != 0) { |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk("PPR with IU_REQ outstanding\n"); |
| #endif |
| ahd->msg_flags |= MSG_FLAG_EXPECT_PPR_BUSFREE; |
| } |
| } |
| } |
| |
| update_needed += ahd_update_neg_request(ahd, devinfo, tstate, |
| tinfo, AHD_NEG_TO_GOAL); |
| |
| if (update_needed && active) |
| ahd_update_pending_scbs(ahd); |
| } |
| |
| /* |
| * Update the user/goal/curr tables of wide negotiation |
| * parameters as well as, in the case of a current or active update, |
| * any data structures on the host controller. In the case of an |
| * active update, the specified target is currently talking to us on |
| * the bus, so the transfer parameter update must take effect |
| * immediately. |
| */ |
| void |
| ahd_set_width(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| u_int width, u_int type, int paused) |
| { |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_tmode_tstate *tstate; |
| u_int oldwidth; |
| int active; |
| int update_needed; |
| |
| active = (type & AHD_TRANS_ACTIVE) == AHD_TRANS_ACTIVE; |
| update_needed = 0; |
| tinfo = ahd_fetch_transinfo(ahd, devinfo->channel, devinfo->our_scsiid, |
| devinfo->target, &tstate); |
| |
| if ((type & AHD_TRANS_USER) != 0) |
| tinfo->user.width = width; |
| |
| if ((type & AHD_TRANS_GOAL) != 0) |
| tinfo->goal.width = width; |
| |
| oldwidth = tinfo->curr.width; |
| if ((type & AHD_TRANS_CUR) != 0 && oldwidth != width) { |
| |
| update_needed++; |
| |
| tinfo->curr.width = width; |
| ahd_send_async(ahd, devinfo->channel, devinfo->target, |
| CAM_LUN_WILDCARD, AC_TRANSFER_NEG); |
| if (bootverbose) { |
| printk("%s: target %d using %dbit transfers\n", |
| ahd_name(ahd), devinfo->target, |
| 8 * (0x01 << width)); |
| } |
| } |
| |
| if ((type & AHD_TRANS_CUR) != 0) { |
| if (!paused) |
| ahd_pause(ahd); |
| ahd_update_neg_table(ahd, devinfo, &tinfo->curr); |
| if (!paused) |
| ahd_unpause(ahd); |
| } |
| |
| update_needed += ahd_update_neg_request(ahd, devinfo, tstate, |
| tinfo, AHD_NEG_TO_GOAL); |
| if (update_needed && active) |
| ahd_update_pending_scbs(ahd); |
| |
| } |
| |
| /* |
| * Update the current state of tagged queuing for a given target. |
| */ |
| static void |
| ahd_set_tags(struct ahd_softc *ahd, struct scsi_cmnd *cmd, |
| struct ahd_devinfo *devinfo, ahd_queue_alg alg) |
| { |
| struct scsi_device *sdev = cmd->device; |
| |
| ahd_platform_set_tags(ahd, sdev, devinfo, alg); |
| ahd_send_async(ahd, devinfo->channel, devinfo->target, |
| devinfo->lun, AC_TRANSFER_NEG); |
| } |
| |
| static void |
| ahd_update_neg_table(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| struct ahd_transinfo *tinfo) |
| { |
| ahd_mode_state saved_modes; |
| u_int period; |
| u_int ppr_opts; |
| u_int con_opts; |
| u_int offset; |
| u_int saved_negoaddr; |
| uint8_t iocell_opts[sizeof(ahd->iocell_opts)]; |
| |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| |
| saved_negoaddr = ahd_inb(ahd, NEGOADDR); |
| ahd_outb(ahd, NEGOADDR, devinfo->target); |
| period = tinfo->period; |
| offset = tinfo->offset; |
| memcpy(iocell_opts, ahd->iocell_opts, sizeof(ahd->iocell_opts)); |
| ppr_opts = tinfo->ppr_options & (MSG_EXT_PPR_QAS_REQ|MSG_EXT_PPR_DT_REQ |
| |MSG_EXT_PPR_IU_REQ|MSG_EXT_PPR_RTI); |
| con_opts = 0; |
| if (period == 0) |
| period = AHD_SYNCRATE_ASYNC; |
| if (period == AHD_SYNCRATE_160) { |
| |
| if ((ahd->bugs & AHD_PACED_NEGTABLE_BUG) != 0) { |
| /* |
| * When the SPI4 spec was finalized, PACE transfers |
| * was not made a configurable option in the PPR |
| * message. Instead it is assumed to be enabled for |
| * any syncrate faster than 80MHz. Nevertheless, |
| * Harpoon2A4 allows this to be configurable. |
| * |
| * Harpoon2A4 also assumes at most 2 data bytes per |
| * negotiated REQ/ACK offset. Paced transfers take |
| * 4, so we must adjust our offset. |
| */ |
| ppr_opts |= PPROPT_PACE; |
| offset *= 2; |
| |
| /* |
| * Harpoon2A assumed that there would be a |
| * fallback rate between 160MHz and 80MHz, |
| * so 7 is used as the period factor rather |
| * than 8 for 160MHz. |
| */ |
| period = AHD_SYNCRATE_REVA_160; |
| } |
| if ((tinfo->ppr_options & MSG_EXT_PPR_PCOMP_EN) == 0) |
| iocell_opts[AHD_PRECOMP_SLEW_INDEX] &= |
| ~AHD_PRECOMP_MASK; |
| } else { |
| /* |
| * Precomp should be disabled for non-paced transfers. |
| */ |
| iocell_opts[AHD_PRECOMP_SLEW_INDEX] &= ~AHD_PRECOMP_MASK; |
| |
| if ((ahd->features & AHD_NEW_IOCELL_OPTS) != 0 |
| && (ppr_opts & MSG_EXT_PPR_DT_REQ) != 0 |
| && (ppr_opts & MSG_EXT_PPR_IU_REQ) == 0) { |
| /* |
| * Slow down our CRC interval to be |
| * compatible with non-packetized |
| * U160 devices that can't handle a |
| * CRC at full speed. |
| */ |
| con_opts |= ENSLOWCRC; |
| } |
| |
| if ((ahd->bugs & AHD_PACED_NEGTABLE_BUG) != 0) { |
| /* |
| * On H2A4, revert to a slower slewrate |
| * on non-paced transfers. |
| */ |
| iocell_opts[AHD_PRECOMP_SLEW_INDEX] &= |
| ~AHD_SLEWRATE_MASK; |
| } |
| } |
| |
| ahd_outb(ahd, ANNEXCOL, AHD_ANNEXCOL_PRECOMP_SLEW); |
| ahd_outb(ahd, ANNEXDAT, iocell_opts[AHD_PRECOMP_SLEW_INDEX]); |
| ahd_outb(ahd, ANNEXCOL, AHD_ANNEXCOL_AMPLITUDE); |
| ahd_outb(ahd, ANNEXDAT, iocell_opts[AHD_AMPLITUDE_INDEX]); |
| |
| ahd_outb(ahd, NEGPERIOD, period); |
| ahd_outb(ahd, NEGPPROPTS, ppr_opts); |
| ahd_outb(ahd, NEGOFFSET, offset); |
| |
| if (tinfo->width == MSG_EXT_WDTR_BUS_16_BIT) |
| con_opts |= WIDEXFER; |
| |
| /* |
| * Slow down our CRC interval to be |
| * compatible with packetized U320 devices |
| * that can't handle a CRC at full speed |
| */ |
| if (ahd->features & AHD_AIC79XXB_SLOWCRC) { |
| con_opts |= ENSLOWCRC; |
| } |
| |
| /* |
| * During packetized transfers, the target will |
| * give us the opportunity to send command packets |
| * without us asserting attention. |
| */ |
| if ((tinfo->ppr_options & MSG_EXT_PPR_IU_REQ) == 0) |
| con_opts |= ENAUTOATNO; |
| ahd_outb(ahd, NEGCONOPTS, con_opts); |
| ahd_outb(ahd, NEGOADDR, saved_negoaddr); |
| ahd_restore_modes(ahd, saved_modes); |
| } |
| |
| /* |
| * When the transfer settings for a connection change, setup for |
| * negotiation in pending SCBs to effect the change as quickly as |
| * possible. We also cancel any negotiations that are scheduled |
| * for inflight SCBs that have not been started yet. |
| */ |
| static void |
| ahd_update_pending_scbs(struct ahd_softc *ahd) |
| { |
| struct scb *pending_scb; |
| int pending_scb_count; |
| int paused; |
| u_int saved_scbptr; |
| ahd_mode_state saved_modes; |
| |
| /* |
| * Traverse the pending SCB list and ensure that all of the |
| * SCBs there have the proper settings. We can only safely |
| * clear the negotiation required flag (setting requires the |
| * execution queue to be modified) and this is only possible |
| * if we are not already attempting to select out for this |
| * SCB. For this reason, all callers only call this routine |
| * if we are changing the negotiation settings for the currently |
| * active transaction on the bus. |
| */ |
| pending_scb_count = 0; |
| LIST_FOREACH(pending_scb, &ahd->pending_scbs, pending_links) { |
| struct ahd_devinfo devinfo; |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_tmode_tstate *tstate; |
| |
| ahd_scb_devinfo(ahd, &devinfo, pending_scb); |
| tinfo = ahd_fetch_transinfo(ahd, devinfo.channel, |
| devinfo.our_scsiid, |
| devinfo.target, &tstate); |
| if ((tstate->auto_negotiate & devinfo.target_mask) == 0 |
| && (pending_scb->flags & SCB_AUTO_NEGOTIATE) != 0) { |
| pending_scb->flags &= ~SCB_AUTO_NEGOTIATE; |
| pending_scb->hscb->control &= ~MK_MESSAGE; |
| } |
| ahd_sync_scb(ahd, pending_scb, |
| BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); |
| pending_scb_count++; |
| } |
| |
| if (pending_scb_count == 0) |
| return; |
| |
| if (ahd_is_paused(ahd)) { |
| paused = 1; |
| } else { |
| paused = 0; |
| ahd_pause(ahd); |
| } |
| |
| /* |
| * Force the sequencer to reinitialize the selection for |
| * the command at the head of the execution queue if it |
| * has already been setup. The negotiation changes may |
| * effect whether we select-out with ATN. It is only |
| * safe to clear ENSELO when the bus is not free and no |
| * selection is in progres or completed. |
| */ |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| if ((ahd_inb(ahd, SCSISIGI) & BSYI) != 0 |
| && (ahd_inb(ahd, SSTAT0) & (SELDO|SELINGO)) == 0) |
| ahd_outb(ahd, SCSISEQ0, ahd_inb(ahd, SCSISEQ0) & ~ENSELO); |
| saved_scbptr = ahd_get_scbptr(ahd); |
| /* Ensure that the hscbs down on the card match the new information */ |
| LIST_FOREACH(pending_scb, &ahd->pending_scbs, pending_links) { |
| u_int scb_tag; |
| u_int control; |
| |
| scb_tag = SCB_GET_TAG(pending_scb); |
| ahd_set_scbptr(ahd, scb_tag); |
| control = ahd_inb_scbram(ahd, SCB_CONTROL); |
| control &= ~MK_MESSAGE; |
| control |= pending_scb->hscb->control & MK_MESSAGE; |
| ahd_outb(ahd, SCB_CONTROL, control); |
| } |
| ahd_set_scbptr(ahd, saved_scbptr); |
| ahd_restore_modes(ahd, saved_modes); |
| |
| if (paused == 0) |
| ahd_unpause(ahd); |
| } |
| |
| /**************************** Pathing Information *****************************/ |
| static void |
| ahd_fetch_devinfo(struct ahd_softc *ahd, struct ahd_devinfo *devinfo) |
| { |
| ahd_mode_state saved_modes; |
| u_int saved_scsiid; |
| role_t role; |
| int our_id; |
| |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| |
| if (ahd_inb(ahd, SSTAT0) & TARGET) |
| role = ROLE_TARGET; |
| else |
| role = ROLE_INITIATOR; |
| |
| if (role == ROLE_TARGET |
| && (ahd_inb(ahd, SEQ_FLAGS) & CMDPHASE_PENDING) != 0) { |
| /* We were selected, so pull our id from TARGIDIN */ |
| our_id = ahd_inb(ahd, TARGIDIN) & OID; |
| } else if (role == ROLE_TARGET) |
| our_id = ahd_inb(ahd, TOWNID); |
| else |
| our_id = ahd_inb(ahd, IOWNID); |
| |
| saved_scsiid = ahd_inb(ahd, SAVED_SCSIID); |
| ahd_compile_devinfo(devinfo, |
| our_id, |
| SCSIID_TARGET(ahd, saved_scsiid), |
| ahd_inb(ahd, SAVED_LUN), |
| SCSIID_CHANNEL(ahd, saved_scsiid), |
| role); |
| ahd_restore_modes(ahd, saved_modes); |
| } |
| |
| void |
| ahd_print_devinfo(struct ahd_softc *ahd, struct ahd_devinfo *devinfo) |
| { |
| printk("%s:%c:%d:%d: ", ahd_name(ahd), 'A', |
| devinfo->target, devinfo->lun); |
| } |
| |
| static const struct ahd_phase_table_entry* |
| ahd_lookup_phase_entry(int phase) |
| { |
| const struct ahd_phase_table_entry *entry; |
| const struct ahd_phase_table_entry *last_entry; |
| |
| /* |
| * num_phases doesn't include the default entry which |
| * will be returned if the phase doesn't match. |
| */ |
| last_entry = &ahd_phase_table[num_phases]; |
| for (entry = ahd_phase_table; entry < last_entry; entry++) { |
| if (phase == entry->phase) |
| break; |
| } |
| return (entry); |
| } |
| |
| void |
| ahd_compile_devinfo(struct ahd_devinfo *devinfo, u_int our_id, u_int target, |
| u_int lun, char channel, role_t role) |
| { |
| devinfo->our_scsiid = our_id; |
| devinfo->target = target; |
| devinfo->lun = lun; |
| devinfo->target_offset = target; |
| devinfo->channel = channel; |
| devinfo->role = role; |
| if (channel == 'B') |
| devinfo->target_offset += 8; |
| devinfo->target_mask = (0x01 << devinfo->target_offset); |
| } |
| |
| static void |
| ahd_scb_devinfo(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| struct scb *scb) |
| { |
| role_t role; |
| int our_id; |
| |
| our_id = SCSIID_OUR_ID(scb->hscb->scsiid); |
| role = ROLE_INITIATOR; |
| if ((scb->hscb->control & TARGET_SCB) != 0) |
| role = ROLE_TARGET; |
| ahd_compile_devinfo(devinfo, our_id, SCB_GET_TARGET(ahd, scb), |
| SCB_GET_LUN(scb), SCB_GET_CHANNEL(ahd, scb), role); |
| } |
| |
| |
| /************************ Message Phase Processing ****************************/ |
| /* |
| * When an initiator transaction with the MK_MESSAGE flag either reconnects |
| * or enters the initial message out phase, we are interrupted. Fill our |
| * outgoing message buffer with the appropriate message and beging handing |
| * the message phase(s) manually. |
| */ |
| static void |
| ahd_setup_initiator_msgout(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| struct scb *scb) |
| { |
| /* |
| * To facilitate adding multiple messages together, |
| * each routine should increment the index and len |
| * variables instead of setting them explicitly. |
| */ |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 0; |
| |
| if (ahd_currently_packetized(ahd)) |
| ahd->msg_flags |= MSG_FLAG_PACKETIZED; |
| |
| if (ahd->send_msg_perror |
| && ahd_inb(ahd, MSG_OUT) == HOST_MSG) { |
| ahd->msgout_buf[ahd->msgout_index++] = ahd->send_msg_perror; |
| ahd->msgout_len++; |
| ahd->msg_type = MSG_TYPE_INITIATOR_MSGOUT; |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk("Setting up for Parity Error delivery\n"); |
| #endif |
| return; |
| } else if (scb == NULL) { |
| printk("%s: WARNING. No pending message for " |
| "I_T msgin. Issuing NO-OP\n", ahd_name(ahd)); |
| ahd->msgout_buf[ahd->msgout_index++] = MSG_NOOP; |
| ahd->msgout_len++; |
| ahd->msg_type = MSG_TYPE_INITIATOR_MSGOUT; |
| return; |
| } |
| |
| if ((scb->flags & SCB_DEVICE_RESET) == 0 |
| && (scb->flags & SCB_PACKETIZED) == 0 |
| && ahd_inb(ahd, MSG_OUT) == MSG_IDENTIFYFLAG) { |
| u_int identify_msg; |
| |
| identify_msg = MSG_IDENTIFYFLAG | SCB_GET_LUN(scb); |
| if ((scb->hscb->control & DISCENB) != 0) |
| identify_msg |= MSG_IDENTIFY_DISCFLAG; |
| ahd->msgout_buf[ahd->msgout_index++] = identify_msg; |
| ahd->msgout_len++; |
| |
| if ((scb->hscb->control & TAG_ENB) != 0) { |
| ahd->msgout_buf[ahd->msgout_index++] = |
| scb->hscb->control & (TAG_ENB|SCB_TAG_TYPE); |
| ahd->msgout_buf[ahd->msgout_index++] = SCB_GET_TAG(scb); |
| ahd->msgout_len += 2; |
| } |
| } |
| |
| if (scb->flags & SCB_DEVICE_RESET) { |
| ahd->msgout_buf[ahd->msgout_index++] = MSG_BUS_DEV_RESET; |
| ahd->msgout_len++; |
| ahd_print_path(ahd, scb); |
| printk("Bus Device Reset Message Sent\n"); |
| /* |
| * Clear our selection hardware in advance of |
| * the busfree. We may have an entry in the waiting |
| * Q for this target, and we don't want to go about |
| * selecting while we handle the busfree and blow it |
| * away. |
| */ |
| ahd_outb(ahd, SCSISEQ0, 0); |
| } else if ((scb->flags & SCB_ABORT) != 0) { |
| |
| if ((scb->hscb->control & TAG_ENB) != 0) { |
| ahd->msgout_buf[ahd->msgout_index++] = MSG_ABORT_TAG; |
| } else { |
| ahd->msgout_buf[ahd->msgout_index++] = MSG_ABORT; |
| } |
| ahd->msgout_len++; |
| ahd_print_path(ahd, scb); |
| printk("Abort%s Message Sent\n", |
| (scb->hscb->control & TAG_ENB) != 0 ? " Tag" : ""); |
| /* |
| * Clear our selection hardware in advance of |
| * the busfree. We may have an entry in the waiting |
| * Q for this target, and we don't want to go about |
| * selecting while we handle the busfree and blow it |
| * away. |
| */ |
| ahd_outb(ahd, SCSISEQ0, 0); |
| } else if ((scb->flags & (SCB_AUTO_NEGOTIATE|SCB_NEGOTIATE)) != 0) { |
| ahd_build_transfer_msg(ahd, devinfo); |
| /* |
| * Clear our selection hardware in advance of potential |
| * PPR IU status change busfree. We may have an entry in |
| * the waiting Q for this target, and we don't want to go |
| * about selecting while we handle the busfree and blow |
| * it away. |
| */ |
| ahd_outb(ahd, SCSISEQ0, 0); |
| } else { |
| printk("ahd_intr: AWAITING_MSG for an SCB that " |
| "does not have a waiting message\n"); |
| printk("SCSIID = %x, target_mask = %x\n", scb->hscb->scsiid, |
| devinfo->target_mask); |
| panic("SCB = %d, SCB Control = %x:%x, MSG_OUT = %x " |
| "SCB flags = %x", SCB_GET_TAG(scb), scb->hscb->control, |
| ahd_inb_scbram(ahd, SCB_CONTROL), ahd_inb(ahd, MSG_OUT), |
| scb->flags); |
| } |
| |
| /* |
| * Clear the MK_MESSAGE flag from the SCB so we aren't |
| * asked to send this message again. |
| */ |
| ahd_outb(ahd, SCB_CONTROL, |
| ahd_inb_scbram(ahd, SCB_CONTROL) & ~MK_MESSAGE); |
| scb->hscb->control &= ~MK_MESSAGE; |
| ahd->msgout_index = 0; |
| ahd->msg_type = MSG_TYPE_INITIATOR_MSGOUT; |
| } |
| |
| /* |
| * Build an appropriate transfer negotiation message for the |
| * currently active target. |
| */ |
| static void |
| ahd_build_transfer_msg(struct ahd_softc *ahd, struct ahd_devinfo *devinfo) |
| { |
| /* |
| * We need to initiate transfer negotiations. |
| * If our current and goal settings are identical, |
| * we want to renegotiate due to a check condition. |
| */ |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_tmode_tstate *tstate; |
| int dowide; |
| int dosync; |
| int doppr; |
| u_int period; |
| u_int ppr_options; |
| u_int offset; |
| |
| tinfo = ahd_fetch_transinfo(ahd, devinfo->channel, devinfo->our_scsiid, |
| devinfo->target, &tstate); |
| /* |
| * Filter our period based on the current connection. |
| * If we can't perform DT transfers on this segment (not in LVD |
| * mode for instance), then our decision to issue a PPR message |
| * may change. |
| */ |
| period = tinfo->goal.period; |
| offset = tinfo->goal.offset; |
| ppr_options = tinfo->goal.ppr_options; |
| /* Target initiated PPR is not allowed in the SCSI spec */ |
| if (devinfo->role == ROLE_TARGET) |
| ppr_options = 0; |
| ahd_devlimited_syncrate(ahd, tinfo, &period, |
| &ppr_options, devinfo->role); |
| dowide = tinfo->curr.width != tinfo->goal.width; |
| dosync = tinfo->curr.offset != offset || tinfo->curr.period != period; |
| /* |
| * Only use PPR if we have options that need it, even if the device |
| * claims to support it. There might be an expander in the way |
| * that doesn't. |
| */ |
| doppr = ppr_options != 0; |
| |
| if (!dowide && !dosync && !doppr) { |
| dowide = tinfo->goal.width != MSG_EXT_WDTR_BUS_8_BIT; |
| dosync = tinfo->goal.offset != 0; |
| } |
| |
| if (!dowide && !dosync && !doppr) { |
| /* |
| * Force async with a WDTR message if we have a wide bus, |
| * or just issue an SDTR with a 0 offset. |
| */ |
| if ((ahd->features & AHD_WIDE) != 0) |
| dowide = 1; |
| else |
| dosync = 1; |
| |
| if (bootverbose) { |
| ahd_print_devinfo(ahd, devinfo); |
| printk("Ensuring async\n"); |
| } |
| } |
| /* Target initiated PPR is not allowed in the SCSI spec */ |
| if (devinfo->role == ROLE_TARGET) |
| doppr = 0; |
| |
| /* |
| * Both the PPR message and SDTR message require the |
| * goal syncrate to be limited to what the target device |
| * is capable of handling (based on whether an LVD->SE |
| * expander is on the bus), so combine these two cases. |
| * Regardless, guarantee that if we are using WDTR and SDTR |
| * messages that WDTR comes first. |
| */ |
| if (doppr || (dosync && !dowide)) { |
| |
| offset = tinfo->goal.offset; |
| ahd_validate_offset(ahd, tinfo, period, &offset, |
| doppr ? tinfo->goal.width |
| : tinfo->curr.width, |
| devinfo->role); |
| if (doppr) { |
| ahd_construct_ppr(ahd, devinfo, period, offset, |
| tinfo->goal.width, ppr_options); |
| } else { |
| ahd_construct_sdtr(ahd, devinfo, period, offset); |
| } |
| } else { |
| ahd_construct_wdtr(ahd, devinfo, tinfo->goal.width); |
| } |
| } |
| |
| /* |
| * Build a synchronous negotiation message in our message |
| * buffer based on the input parameters. |
| */ |
| static void |
| ahd_construct_sdtr(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| u_int period, u_int offset) |
| { |
| if (offset == 0) |
| period = AHD_ASYNC_XFER_PERIOD; |
| ahd->msgout_index += spi_populate_sync_msg( |
| ahd->msgout_buf + ahd->msgout_index, period, offset); |
| ahd->msgout_len += 5; |
| if (bootverbose) { |
| printk("(%s:%c:%d:%d): Sending SDTR period %x, offset %x\n", |
| ahd_name(ahd), devinfo->channel, devinfo->target, |
| devinfo->lun, period, offset); |
| } |
| } |
| |
| /* |
| * Build a wide negotiateion message in our message |
| * buffer based on the input parameters. |
| */ |
| static void |
| ahd_construct_wdtr(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| u_int bus_width) |
| { |
| ahd->msgout_index += spi_populate_width_msg( |
| ahd->msgout_buf + ahd->msgout_index, bus_width); |
| ahd->msgout_len += 4; |
| if (bootverbose) { |
| printk("(%s:%c:%d:%d): Sending WDTR %x\n", |
| ahd_name(ahd), devinfo->channel, devinfo->target, |
| devinfo->lun, bus_width); |
| } |
| } |
| |
| /* |
| * Build a parallel protocol request message in our message |
| * buffer based on the input parameters. |
| */ |
| static void |
| ahd_construct_ppr(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| u_int period, u_int offset, u_int bus_width, |
| u_int ppr_options) |
| { |
| /* |
| * Always request precompensation from |
| * the other target if we are running |
| * at paced syncrates. |
| */ |
| if (period <= AHD_SYNCRATE_PACED) |
| ppr_options |= MSG_EXT_PPR_PCOMP_EN; |
| if (offset == 0) |
| period = AHD_ASYNC_XFER_PERIOD; |
| ahd->msgout_index += spi_populate_ppr_msg( |
| ahd->msgout_buf + ahd->msgout_index, period, offset, |
| bus_width, ppr_options); |
| ahd->msgout_len += 8; |
| if (bootverbose) { |
| printk("(%s:%c:%d:%d): Sending PPR bus_width %x, period %x, " |
| "offset %x, ppr_options %x\n", ahd_name(ahd), |
| devinfo->channel, devinfo->target, devinfo->lun, |
| bus_width, period, offset, ppr_options); |
| } |
| } |
| |
| /* |
| * Clear any active message state. |
| */ |
| static void |
| ahd_clear_msg_state(struct ahd_softc *ahd) |
| { |
| ahd_mode_state saved_modes; |
| |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| ahd->send_msg_perror = 0; |
| ahd->msg_flags = MSG_FLAG_NONE; |
| ahd->msgout_len = 0; |
| ahd->msgin_index = 0; |
| ahd->msg_type = MSG_TYPE_NONE; |
| if ((ahd_inb(ahd, SCSISIGO) & ATNO) != 0) { |
| /* |
| * The target didn't care to respond to our |
| * message request, so clear ATN. |
| */ |
| ahd_outb(ahd, CLRSINT1, CLRATNO); |
| } |
| ahd_outb(ahd, MSG_OUT, MSG_NOOP); |
| ahd_outb(ahd, SEQ_FLAGS2, |
| ahd_inb(ahd, SEQ_FLAGS2) & ~TARGET_MSG_PENDING); |
| ahd_restore_modes(ahd, saved_modes); |
| } |
| |
| /* |
| * Manual message loop handler. |
| */ |
| static void |
| ahd_handle_message_phase(struct ahd_softc *ahd) |
| { |
| struct ahd_devinfo devinfo; |
| u_int bus_phase; |
| int end_session; |
| |
| ahd_fetch_devinfo(ahd, &devinfo); |
| end_session = FALSE; |
| bus_phase = ahd_inb(ahd, LASTPHASE); |
| |
| if ((ahd_inb(ahd, LQISTAT2) & LQIPHASE_OUTPKT) != 0) { |
| printk("LQIRETRY for LQIPHASE_OUTPKT\n"); |
| ahd_outb(ahd, LQCTL2, LQIRETRY); |
| } |
| reswitch: |
| switch (ahd->msg_type) { |
| case MSG_TYPE_INITIATOR_MSGOUT: |
| { |
| int lastbyte; |
| int phasemis; |
| int msgdone; |
| |
| if (ahd->msgout_len == 0 && ahd->send_msg_perror == 0) |
| panic("HOST_MSG_LOOP interrupt with no active message"); |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) { |
| ahd_print_devinfo(ahd, &devinfo); |
| printk("INITIATOR_MSG_OUT"); |
| } |
| #endif |
| phasemis = bus_phase != P_MESGOUT; |
| if (phasemis) { |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) { |
| printk(" PHASEMIS %s\n", |
| ahd_lookup_phase_entry(bus_phase) |
| ->phasemsg); |
| } |
| #endif |
| if (bus_phase == P_MESGIN) { |
| /* |
| * Change gears and see if |
| * this messages is of interest to |
| * us or should be passed back to |
| * the sequencer. |
| */ |
| ahd_outb(ahd, CLRSINT1, CLRATNO); |
| ahd->send_msg_perror = 0; |
| ahd->msg_type = MSG_TYPE_INITIATOR_MSGIN; |
| ahd->msgin_index = 0; |
| goto reswitch; |
| } |
| end_session = TRUE; |
| break; |
| } |
| |
| if (ahd->send_msg_perror) { |
| ahd_outb(ahd, CLRSINT1, CLRATNO); |
| ahd_outb(ahd, CLRSINT1, CLRREQINIT); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk(" byte 0x%x\n", ahd->send_msg_perror); |
| #endif |
| /* |
| * If we are notifying the target of a CRC error |
| * during packetized operations, the target is |
| * within its rights to acknowledge our message |
| * with a busfree. |
| */ |
| if ((ahd->msg_flags & MSG_FLAG_PACKETIZED) != 0 |
| && ahd->send_msg_perror == MSG_INITIATOR_DET_ERR) |
| ahd->msg_flags |= MSG_FLAG_EXPECT_IDE_BUSFREE; |
| |
| ahd_outb(ahd, RETURN_2, ahd->send_msg_perror); |
| ahd_outb(ahd, RETURN_1, CONT_MSG_LOOP_WRITE); |
| break; |
| } |
| |
| msgdone = ahd->msgout_index == ahd->msgout_len; |
| if (msgdone) { |
| /* |
| * The target has requested a retry. |
| * Re-assert ATN, reset our message index to |
| * 0, and try again. |
| */ |
| ahd->msgout_index = 0; |
| ahd_assert_atn(ahd); |
| } |
| |
| lastbyte = ahd->msgout_index == (ahd->msgout_len - 1); |
| if (lastbyte) { |
| /* Last byte is signified by dropping ATN */ |
| ahd_outb(ahd, CLRSINT1, CLRATNO); |
| } |
| |
| /* |
| * Clear our interrupt status and present |
| * the next byte on the bus. |
| */ |
| ahd_outb(ahd, CLRSINT1, CLRREQINIT); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk(" byte 0x%x\n", |
| ahd->msgout_buf[ahd->msgout_index]); |
| #endif |
| ahd_outb(ahd, RETURN_2, ahd->msgout_buf[ahd->msgout_index++]); |
| ahd_outb(ahd, RETURN_1, CONT_MSG_LOOP_WRITE); |
| break; |
| } |
| case MSG_TYPE_INITIATOR_MSGIN: |
| { |
| int phasemis; |
| int message_done; |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) { |
| ahd_print_devinfo(ahd, &devinfo); |
| printk("INITIATOR_MSG_IN"); |
| } |
| #endif |
| phasemis = bus_phase != P_MESGIN; |
| if (phasemis) { |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) { |
| printk(" PHASEMIS %s\n", |
| ahd_lookup_phase_entry(bus_phase) |
| ->phasemsg); |
| } |
| #endif |
| ahd->msgin_index = 0; |
| if (bus_phase == P_MESGOUT |
| && (ahd->send_msg_perror != 0 |
| || (ahd->msgout_len != 0 |
| && ahd->msgout_index == 0))) { |
| ahd->msg_type = MSG_TYPE_INITIATOR_MSGOUT; |
| goto reswitch; |
| } |
| end_session = TRUE; |
| break; |
| } |
| |
| /* Pull the byte in without acking it */ |
| ahd->msgin_buf[ahd->msgin_index] = ahd_inb(ahd, SCSIBUS); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk(" byte 0x%x\n", |
| ahd->msgin_buf[ahd->msgin_index]); |
| #endif |
| |
| message_done = ahd_parse_msg(ahd, &devinfo); |
| |
| if (message_done) { |
| /* |
| * Clear our incoming message buffer in case there |
| * is another message following this one. |
| */ |
| ahd->msgin_index = 0; |
| |
| /* |
| * If this message illicited a response, |
| * assert ATN so the target takes us to the |
| * message out phase. |
| */ |
| if (ahd->msgout_len != 0) { |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) { |
| ahd_print_devinfo(ahd, &devinfo); |
| printk("Asserting ATN for response\n"); |
| } |
| #endif |
| ahd_assert_atn(ahd); |
| } |
| } else |
| ahd->msgin_index++; |
| |
| if (message_done == MSGLOOP_TERMINATED) { |
| end_session = TRUE; |
| } else { |
| /* Ack the byte */ |
| ahd_outb(ahd, CLRSINT1, CLRREQINIT); |
| ahd_outb(ahd, RETURN_1, CONT_MSG_LOOP_READ); |
| } |
| break; |
| } |
| case MSG_TYPE_TARGET_MSGIN: |
| { |
| int msgdone; |
| int msgout_request; |
| |
| /* |
| * By default, the message loop will continue. |
| */ |
| ahd_outb(ahd, RETURN_1, CONT_MSG_LOOP_TARG); |
| |
| if (ahd->msgout_len == 0) |
| panic("Target MSGIN with no active message"); |
| |
| /* |
| * If we interrupted a mesgout session, the initiator |
| * will not know this until our first REQ. So, we |
| * only honor mesgout requests after we've sent our |
| * first byte. |
| */ |
| if ((ahd_inb(ahd, SCSISIGI) & ATNI) != 0 |
| && ahd->msgout_index > 0) |
| msgout_request = TRUE; |
| else |
| msgout_request = FALSE; |
| |
| if (msgout_request) { |
| |
| /* |
| * Change gears and see if |
| * this messages is of interest to |
| * us or should be passed back to |
| * the sequencer. |
| */ |
| ahd->msg_type = MSG_TYPE_TARGET_MSGOUT; |
| ahd_outb(ahd, SCSISIGO, P_MESGOUT | BSYO); |
| ahd->msgin_index = 0; |
| /* Dummy read to REQ for first byte */ |
| ahd_inb(ahd, SCSIDAT); |
| ahd_outb(ahd, SXFRCTL0, |
| ahd_inb(ahd, SXFRCTL0) | SPIOEN); |
| break; |
| } |
| |
| msgdone = ahd->msgout_index == ahd->msgout_len; |
| if (msgdone) { |
| ahd_outb(ahd, SXFRCTL0, |
| ahd_inb(ahd, SXFRCTL0) & ~SPIOEN); |
| end_session = TRUE; |
| break; |
| } |
| |
| /* |
| * Present the next byte on the bus. |
| */ |
| ahd_outb(ahd, SXFRCTL0, ahd_inb(ahd, SXFRCTL0) | SPIOEN); |
| ahd_outb(ahd, SCSIDAT, ahd->msgout_buf[ahd->msgout_index++]); |
| break; |
| } |
| case MSG_TYPE_TARGET_MSGOUT: |
| { |
| int lastbyte; |
| int msgdone; |
| |
| /* |
| * By default, the message loop will continue. |
| */ |
| ahd_outb(ahd, RETURN_1, CONT_MSG_LOOP_TARG); |
| |
| /* |
| * The initiator signals that this is |
| * the last byte by dropping ATN. |
| */ |
| lastbyte = (ahd_inb(ahd, SCSISIGI) & ATNI) == 0; |
| |
| /* |
| * Read the latched byte, but turn off SPIOEN first |
| * so that we don't inadvertently cause a REQ for the |
| * next byte. |
| */ |
| ahd_outb(ahd, SXFRCTL0, ahd_inb(ahd, SXFRCTL0) & ~SPIOEN); |
| ahd->msgin_buf[ahd->msgin_index] = ahd_inb(ahd, SCSIDAT); |
| msgdone = ahd_parse_msg(ahd, &devinfo); |
| if (msgdone == MSGLOOP_TERMINATED) { |
| /* |
| * The message is *really* done in that it caused |
| * us to go to bus free. The sequencer has already |
| * been reset at this point, so pull the ejection |
| * handle. |
| */ |
| return; |
| } |
| |
| ahd->msgin_index++; |
| |
| /* |
| * XXX Read spec about initiator dropping ATN too soon |
| * and use msgdone to detect it. |
| */ |
| if (msgdone == MSGLOOP_MSGCOMPLETE) { |
| ahd->msgin_index = 0; |
| |
| /* |
| * If this message illicited a response, transition |
| * to the Message in phase and send it. |
| */ |
| if (ahd->msgout_len != 0) { |
| ahd_outb(ahd, SCSISIGO, P_MESGIN | BSYO); |
| ahd_outb(ahd, SXFRCTL0, |
| ahd_inb(ahd, SXFRCTL0) | SPIOEN); |
| ahd->msg_type = MSG_TYPE_TARGET_MSGIN; |
| ahd->msgin_index = 0; |
| break; |
| } |
| } |
| |
| if (lastbyte) |
| end_session = TRUE; |
| else { |
| /* Ask for the next byte. */ |
| ahd_outb(ahd, SXFRCTL0, |
| ahd_inb(ahd, SXFRCTL0) | SPIOEN); |
| } |
| |
| break; |
| } |
| default: |
| panic("Unknown REQINIT message type"); |
| } |
| |
| if (end_session) { |
| if ((ahd->msg_flags & MSG_FLAG_PACKETIZED) != 0) { |
| printk("%s: Returning to Idle Loop\n", |
| ahd_name(ahd)); |
| ahd_clear_msg_state(ahd); |
| |
| /* |
| * Perform the equivalent of a clear_target_state. |
| */ |
| ahd_outb(ahd, LASTPHASE, P_BUSFREE); |
| ahd_outb(ahd, SEQ_FLAGS, NOT_IDENTIFIED|NO_CDB_SENT); |
| ahd_outb(ahd, SEQCTL0, FASTMODE|SEQRESET); |
| } else { |
| ahd_clear_msg_state(ahd); |
| ahd_outb(ahd, RETURN_1, EXIT_MSG_LOOP); |
| } |
| } |
| } |
| |
| /* |
| * See if we sent a particular extended message to the target. |
| * If "full" is true, return true only if the target saw the full |
| * message. If "full" is false, return true if the target saw at |
| * least the first byte of the message. |
| */ |
| static int |
| ahd_sent_msg(struct ahd_softc *ahd, ahd_msgtype type, u_int msgval, int full) |
| { |
| int found; |
| u_int index; |
| |
| found = FALSE; |
| index = 0; |
| |
| while (index < ahd->msgout_len) { |
| if (ahd->msgout_buf[index] == MSG_EXTENDED) { |
| u_int end_index; |
| |
| end_index = index + 1 + ahd->msgout_buf[index + 1]; |
| if (ahd->msgout_buf[index+2] == msgval |
| && type == AHDMSG_EXT) { |
| |
| if (full) { |
| if (ahd->msgout_index > end_index) |
| found = TRUE; |
| } else if (ahd->msgout_index > index) |
| found = TRUE; |
| } |
| index = end_index; |
| } else if (ahd->msgout_buf[index] >= MSG_SIMPLE_TASK |
| && ahd->msgout_buf[index] <= MSG_IGN_WIDE_RESIDUE) { |
| |
| /* Skip tag type and tag id or residue param*/ |
| index += 2; |
| } else { |
| /* Single byte message */ |
| if (type == AHDMSG_1B |
| && ahd->msgout_index > index |
| && (ahd->msgout_buf[index] == msgval |
| || ((ahd->msgout_buf[index] & MSG_IDENTIFYFLAG) != 0 |
| && msgval == MSG_IDENTIFYFLAG))) |
| found = TRUE; |
| index++; |
| } |
| |
| if (found) |
| break; |
| } |
| return (found); |
| } |
| |
| /* |
| * Wait for a complete incoming message, parse it, and respond accordingly. |
| */ |
| static int |
| ahd_parse_msg(struct ahd_softc *ahd, struct ahd_devinfo *devinfo) |
| { |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_tmode_tstate *tstate; |
| int reject; |
| int done; |
| int response; |
| |
| done = MSGLOOP_IN_PROG; |
| response = FALSE; |
| reject = FALSE; |
| tinfo = ahd_fetch_transinfo(ahd, devinfo->channel, devinfo->our_scsiid, |
| devinfo->target, &tstate); |
| |
| /* |
| * Parse as much of the message as is available, |
| * rejecting it if we don't support it. When |
| * the entire message is available and has been |
| * handled, return MSGLOOP_MSGCOMPLETE, indicating |
| * that we have parsed an entire message. |
| * |
| * In the case of extended messages, we accept the length |
| * byte outright and perform more checking once we know the |
| * extended message type. |
| */ |
| switch (ahd->msgin_buf[0]) { |
| case MSG_DISCONNECT: |
| case MSG_SAVEDATAPOINTER: |
| case MSG_CMDCOMPLETE: |
| case MSG_RESTOREPOINTERS: |
| case MSG_IGN_WIDE_RESIDUE: |
| /* |
| * End our message loop as these are messages |
| * the sequencer handles on its own. |
| */ |
| done = MSGLOOP_TERMINATED; |
| break; |
| case MSG_MESSAGE_REJECT: |
| response = ahd_handle_msg_reject(ahd, devinfo); |
| /* FALLTHROUGH */ |
| case MSG_NOOP: |
| done = MSGLOOP_MSGCOMPLETE; |
| break; |
| case MSG_EXTENDED: |
| { |
| /* Wait for enough of the message to begin validation */ |
| if (ahd->msgin_index < 2) |
| break; |
| switch (ahd->msgin_buf[2]) { |
| case MSG_EXT_SDTR: |
| { |
| u_int period; |
| u_int ppr_options; |
| u_int offset; |
| u_int saved_offset; |
| |
| if (ahd->msgin_buf[1] != MSG_EXT_SDTR_LEN) { |
| reject = TRUE; |
| break; |
| } |
| |
| /* |
| * Wait until we have both args before validating |
| * and acting on this message. |
| * |
| * Add one to MSG_EXT_SDTR_LEN to account for |
| * the extended message preamble. |
| */ |
| if (ahd->msgin_index < (MSG_EXT_SDTR_LEN + 1)) |
| break; |
| |
| period = ahd->msgin_buf[3]; |
| ppr_options = 0; |
| saved_offset = offset = ahd->msgin_buf[4]; |
| ahd_devlimited_syncrate(ahd, tinfo, &period, |
| &ppr_options, devinfo->role); |
| ahd_validate_offset(ahd, tinfo, period, &offset, |
| tinfo->curr.width, devinfo->role); |
| if (bootverbose) { |
| printk("(%s:%c:%d:%d): Received " |
| "SDTR period %x, offset %x\n\t" |
| "Filtered to period %x, offset %x\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun, |
| ahd->msgin_buf[3], saved_offset, |
| period, offset); |
| } |
| ahd_set_syncrate(ahd, devinfo, period, |
| offset, ppr_options, |
| AHD_TRANS_ACTIVE|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| |
| /* |
| * See if we initiated Sync Negotiation |
| * and didn't have to fall down to async |
| * transfers. |
| */ |
| if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_SDTR, TRUE)) { |
| /* We started it */ |
| if (saved_offset != offset) { |
| /* Went too low - force async */ |
| reject = TRUE; |
| } |
| } else { |
| /* |
| * Send our own SDTR in reply |
| */ |
| if (bootverbose |
| && devinfo->role == ROLE_INITIATOR) { |
| printk("(%s:%c:%d:%d): Target " |
| "Initiated SDTR\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun); |
| } |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 0; |
| ahd_construct_sdtr(ahd, devinfo, |
| period, offset); |
| ahd->msgout_index = 0; |
| response = TRUE; |
| } |
| done = MSGLOOP_MSGCOMPLETE; |
| break; |
| } |
| case MSG_EXT_WDTR: |
| { |
| u_int bus_width; |
| u_int saved_width; |
| u_int sending_reply; |
| |
| sending_reply = FALSE; |
| if (ahd->msgin_buf[1] != MSG_EXT_WDTR_LEN) { |
| reject = TRUE; |
| break; |
| } |
| |
| /* |
| * Wait until we have our arg before validating |
| * and acting on this message. |
| * |
| * Add one to MSG_EXT_WDTR_LEN to account for |
| * the extended message preamble. |
| */ |
| if (ahd->msgin_index < (MSG_EXT_WDTR_LEN + 1)) |
| break; |
| |
| bus_width = ahd->msgin_buf[3]; |
| saved_width = bus_width; |
| ahd_validate_width(ahd, tinfo, &bus_width, |
| devinfo->role); |
| if (bootverbose) { |
| printk("(%s:%c:%d:%d): Received WDTR " |
| "%x filtered to %x\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun, |
| saved_width, bus_width); |
| } |
| |
| if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_WDTR, TRUE)) { |
| /* |
| * Don't send a WDTR back to the |
| * target, since we asked first. |
| * If the width went higher than our |
| * request, reject it. |
| */ |
| if (saved_width > bus_width) { |
| reject = TRUE; |
| printk("(%s:%c:%d:%d): requested %dBit " |
| "transfers. Rejecting...\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun, |
| 8 * (0x01 << bus_width)); |
| bus_width = 0; |
| } |
| } else { |
| /* |
| * Send our own WDTR in reply |
| */ |
| if (bootverbose |
| && devinfo->role == ROLE_INITIATOR) { |
| printk("(%s:%c:%d:%d): Target " |
| "Initiated WDTR\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun); |
| } |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 0; |
| ahd_construct_wdtr(ahd, devinfo, bus_width); |
| ahd->msgout_index = 0; |
| response = TRUE; |
| sending_reply = TRUE; |
| } |
| /* |
| * After a wide message, we are async, but |
| * some devices don't seem to honor this portion |
| * of the spec. Force a renegotiation of the |
| * sync component of our transfer agreement even |
| * if our goal is async. By updating our width |
| * after forcing the negotiation, we avoid |
| * renegotiating for width. |
| */ |
| ahd_update_neg_request(ahd, devinfo, tstate, |
| tinfo, AHD_NEG_ALWAYS); |
| ahd_set_width(ahd, devinfo, bus_width, |
| AHD_TRANS_ACTIVE|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| if (sending_reply == FALSE && reject == FALSE) { |
| |
| /* |
| * We will always have an SDTR to send. |
| */ |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 0; |
| ahd_build_transfer_msg(ahd, devinfo); |
| ahd->msgout_index = 0; |
| response = TRUE; |
| } |
| done = MSGLOOP_MSGCOMPLETE; |
| break; |
| } |
| case MSG_EXT_PPR: |
| { |
| u_int period; |
| u_int offset; |
| u_int bus_width; |
| u_int ppr_options; |
| u_int saved_width; |
| u_int saved_offset; |
| u_int saved_ppr_options; |
| |
| if (ahd->msgin_buf[1] != MSG_EXT_PPR_LEN) { |
| reject = TRUE; |
| break; |
| } |
| |
| /* |
| * Wait until we have all args before validating |
| * and acting on this message. |
| * |
| * Add one to MSG_EXT_PPR_LEN to account for |
| * the extended message preamble. |
| */ |
| if (ahd->msgin_index < (MSG_EXT_PPR_LEN + 1)) |
| break; |
| |
| period = ahd->msgin_buf[3]; |
| offset = ahd->msgin_buf[5]; |
| bus_width = ahd->msgin_buf[6]; |
| saved_width = bus_width; |
| ppr_options = ahd->msgin_buf[7]; |
| /* |
| * According to the spec, a DT only |
| * period factor with no DT option |
| * set implies async. |
| */ |
| if ((ppr_options & MSG_EXT_PPR_DT_REQ) == 0 |
| && period <= 9) |
| offset = 0; |
| saved_ppr_options = ppr_options; |
| saved_offset = offset; |
| |
| /* |
| * Transfer options are only available if we |
| * are negotiating wide. |
| */ |
| if (bus_width == 0) |
| ppr_options &= MSG_EXT_PPR_QAS_REQ; |
| |
| ahd_validate_width(ahd, tinfo, &bus_width, |
| devinfo->role); |
| ahd_devlimited_syncrate(ahd, tinfo, &period, |
| &ppr_options, devinfo->role); |
| ahd_validate_offset(ahd, tinfo, period, &offset, |
| bus_width, devinfo->role); |
| |
| if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_PPR, TRUE)) { |
| /* |
| * If we are unable to do any of the |
| * requested options (we went too low), |
| * then we'll have to reject the message. |
| */ |
| if (saved_width > bus_width |
| || saved_offset != offset |
| || saved_ppr_options != ppr_options) { |
| reject = TRUE; |
| period = 0; |
| offset = 0; |
| bus_width = 0; |
| ppr_options = 0; |
| } |
| } else { |
| if (devinfo->role != ROLE_TARGET) |
| printk("(%s:%c:%d:%d): Target " |
| "Initiated PPR\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun); |
| else |
| printk("(%s:%c:%d:%d): Initiator " |
| "Initiated PPR\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun); |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 0; |
| ahd_construct_ppr(ahd, devinfo, period, offset, |
| bus_width, ppr_options); |
| ahd->msgout_index = 0; |
| response = TRUE; |
| } |
| if (bootverbose) { |
| printk("(%s:%c:%d:%d): Received PPR width %x, " |
| "period %x, offset %x,options %x\n" |
| "\tFiltered to width %x, period %x, " |
| "offset %x, options %x\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun, |
| saved_width, ahd->msgin_buf[3], |
| saved_offset, saved_ppr_options, |
| bus_width, period, offset, ppr_options); |
| } |
| ahd_set_width(ahd, devinfo, bus_width, |
| AHD_TRANS_ACTIVE|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| ahd_set_syncrate(ahd, devinfo, period, |
| offset, ppr_options, |
| AHD_TRANS_ACTIVE|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| |
| done = MSGLOOP_MSGCOMPLETE; |
| break; |
| } |
| default: |
| /* Unknown extended message. Reject it. */ |
| reject = TRUE; |
| break; |
| } |
| break; |
| } |
| #ifdef AHD_TARGET_MODE |
| case MSG_BUS_DEV_RESET: |
| ahd_handle_devreset(ahd, devinfo, CAM_LUN_WILDCARD, |
| CAM_BDR_SENT, |
| "Bus Device Reset Received", |
| /*verbose_level*/0); |
| ahd_restart(ahd); |
| done = MSGLOOP_TERMINATED; |
| break; |
| case MSG_ABORT_TAG: |
| case MSG_ABORT: |
| case MSG_CLEAR_QUEUE: |
| { |
| int tag; |
| |
| /* Target mode messages */ |
| if (devinfo->role != ROLE_TARGET) { |
| reject = TRUE; |
| break; |
| } |
| tag = SCB_LIST_NULL; |
| if (ahd->msgin_buf[0] == MSG_ABORT_TAG) |
| tag = ahd_inb(ahd, INITIATOR_TAG); |
| ahd_abort_scbs(ahd, devinfo->target, devinfo->channel, |
| devinfo->lun, tag, ROLE_TARGET, |
| CAM_REQ_ABORTED); |
| |
| tstate = ahd->enabled_targets[devinfo->our_scsiid]; |
| if (tstate != NULL) { |
| struct ahd_tmode_lstate* lstate; |
| |
| lstate = tstate->enabled_luns[devinfo->lun]; |
| if (lstate != NULL) { |
| ahd_queue_lstate_event(ahd, lstate, |
| devinfo->our_scsiid, |
| ahd->msgin_buf[0], |
| /*arg*/tag); |
| ahd_send_lstate_events(ahd, lstate); |
| } |
| } |
| ahd_restart(ahd); |
| done = MSGLOOP_TERMINATED; |
| break; |
| } |
| #endif |
| case MSG_QAS_REQUEST: |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MESSAGES) != 0) |
| printk("%s: QAS request. SCSISIGI == 0x%x\n", |
| ahd_name(ahd), ahd_inb(ahd, SCSISIGI)); |
| #endif |
| ahd->msg_flags |= MSG_FLAG_EXPECT_QASREJ_BUSFREE; |
| /* FALLTHROUGH */ |
| case MSG_TERM_IO_PROC: |
| default: |
| reject = TRUE; |
| break; |
| } |
| |
| if (reject) { |
| /* |
| * Setup to reject the message. |
| */ |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 1; |
| ahd->msgout_buf[0] = MSG_MESSAGE_REJECT; |
| done = MSGLOOP_MSGCOMPLETE; |
| response = TRUE; |
| } |
| |
| if (done != MSGLOOP_IN_PROG && !response) |
| /* Clear the outgoing message buffer */ |
| ahd->msgout_len = 0; |
| |
| return (done); |
| } |
| |
| /* |
| * Process a message reject message. |
| */ |
| static int |
| ahd_handle_msg_reject(struct ahd_softc *ahd, struct ahd_devinfo *devinfo) |
| { |
| /* |
| * What we care about here is if we had an |
| * outstanding SDTR or WDTR message for this |
| * target. If we did, this is a signal that |
| * the target is refusing negotiation. |
| */ |
| struct scb *scb; |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_tmode_tstate *tstate; |
| u_int scb_index; |
| u_int last_msg; |
| int response = 0; |
| |
| scb_index = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scb_index); |
| tinfo = ahd_fetch_transinfo(ahd, devinfo->channel, |
| devinfo->our_scsiid, |
| devinfo->target, &tstate); |
| /* Might be necessary */ |
| last_msg = ahd_inb(ahd, LAST_MSG); |
| |
| if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_PPR, /*full*/FALSE)) { |
| if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_PPR, /*full*/TRUE) |
| && tinfo->goal.period <= AHD_SYNCRATE_PACED) { |
| /* |
| * Target may not like our SPI-4 PPR Options. |
| * Attempt to negotiate 80MHz which will turn |
| * off these options. |
| */ |
| if (bootverbose) { |
| printk("(%s:%c:%d:%d): PPR Rejected. " |
| "Trying simple U160 PPR\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun); |
| } |
| tinfo->goal.period = AHD_SYNCRATE_DT; |
| tinfo->goal.ppr_options &= MSG_EXT_PPR_IU_REQ |
| | MSG_EXT_PPR_QAS_REQ |
| | MSG_EXT_PPR_DT_REQ; |
| } else { |
| /* |
| * Target does not support the PPR message. |
| * Attempt to negotiate SPI-2 style. |
| */ |
| if (bootverbose) { |
| printk("(%s:%c:%d:%d): PPR Rejected. " |
| "Trying WDTR/SDTR\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun); |
| } |
| tinfo->goal.ppr_options = 0; |
| tinfo->curr.transport_version = 2; |
| tinfo->goal.transport_version = 2; |
| } |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 0; |
| ahd_build_transfer_msg(ahd, devinfo); |
| ahd->msgout_index = 0; |
| response = 1; |
| } else if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_WDTR, /*full*/FALSE)) { |
| |
| /* note 8bit xfers */ |
| printk("(%s:%c:%d:%d): refuses WIDE negotiation. Using " |
| "8bit transfers\n", ahd_name(ahd), |
| devinfo->channel, devinfo->target, devinfo->lun); |
| ahd_set_width(ahd, devinfo, MSG_EXT_WDTR_BUS_8_BIT, |
| AHD_TRANS_ACTIVE|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| /* |
| * No need to clear the sync rate. If the target |
| * did not accept the command, our syncrate is |
| * unaffected. If the target started the negotiation, |
| * but rejected our response, we already cleared the |
| * sync rate before sending our WDTR. |
| */ |
| if (tinfo->goal.offset != tinfo->curr.offset) { |
| |
| /* Start the sync negotiation */ |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 0; |
| ahd_build_transfer_msg(ahd, devinfo); |
| ahd->msgout_index = 0; |
| response = 1; |
| } |
| } else if (ahd_sent_msg(ahd, AHDMSG_EXT, MSG_EXT_SDTR, /*full*/FALSE)) { |
| /* note asynch xfers and clear flag */ |
| ahd_set_syncrate(ahd, devinfo, /*period*/0, |
| /*offset*/0, /*ppr_options*/0, |
| AHD_TRANS_ACTIVE|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| printk("(%s:%c:%d:%d): refuses synchronous negotiation. " |
| "Using asynchronous transfers\n", |
| ahd_name(ahd), devinfo->channel, |
| devinfo->target, devinfo->lun); |
| } else if ((scb->hscb->control & MSG_SIMPLE_TASK) != 0) { |
| int tag_type; |
| int mask; |
| |
| tag_type = (scb->hscb->control & MSG_SIMPLE_TASK); |
| |
| if (tag_type == MSG_SIMPLE_TASK) { |
| printk("(%s:%c:%d:%d): refuses tagged commands. " |
| "Performing non-tagged I/O\n", ahd_name(ahd), |
| devinfo->channel, devinfo->target, devinfo->lun); |
| ahd_set_tags(ahd, scb->io_ctx, devinfo, AHD_QUEUE_NONE); |
| mask = ~0x23; |
| } else { |
| printk("(%s:%c:%d:%d): refuses %s tagged commands. " |
| "Performing simple queue tagged I/O only\n", |
| ahd_name(ahd), devinfo->channel, devinfo->target, |
| devinfo->lun, tag_type == MSG_ORDERED_TASK |
| ? "ordered" : "head of queue"); |
| ahd_set_tags(ahd, scb->io_ctx, devinfo, AHD_QUEUE_BASIC); |
| mask = ~0x03; |
| } |
| |
| /* |
| * Resend the identify for this CCB as the target |
| * may believe that the selection is invalid otherwise. |
| */ |
| ahd_outb(ahd, SCB_CONTROL, |
| ahd_inb_scbram(ahd, SCB_CONTROL) & mask); |
| scb->hscb->control &= mask; |
| ahd_set_transaction_tag(scb, /*enabled*/FALSE, |
| /*type*/MSG_SIMPLE_TASK); |
| ahd_outb(ahd, MSG_OUT, MSG_IDENTIFYFLAG); |
| ahd_assert_atn(ahd); |
| ahd_busy_tcl(ahd, BUILD_TCL(scb->hscb->scsiid, devinfo->lun), |
| SCB_GET_TAG(scb)); |
| |
| /* |
| * Requeue all tagged commands for this target |
| * currently in our possession so they can be |
| * converted to untagged commands. |
| */ |
| ahd_search_qinfifo(ahd, SCB_GET_TARGET(ahd, scb), |
| SCB_GET_CHANNEL(ahd, scb), |
| SCB_GET_LUN(scb), /*tag*/SCB_LIST_NULL, |
| ROLE_INITIATOR, CAM_REQUEUE_REQ, |
| SEARCH_COMPLETE); |
| } else if (ahd_sent_msg(ahd, AHDMSG_1B, MSG_IDENTIFYFLAG, TRUE)) { |
| /* |
| * Most likely the device believes that we had |
| * previously negotiated packetized. |
| */ |
| ahd->msg_flags |= MSG_FLAG_EXPECT_PPR_BUSFREE |
| | MSG_FLAG_IU_REQ_CHANGED; |
| |
| ahd_force_renegotiation(ahd, devinfo); |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 0; |
| ahd_build_transfer_msg(ahd, devinfo); |
| ahd->msgout_index = 0; |
| response = 1; |
| } else { |
| /* |
| * Otherwise, we ignore it. |
| */ |
| printk("%s:%c:%d: Message reject for %x -- ignored\n", |
| ahd_name(ahd), devinfo->channel, devinfo->target, |
| last_msg); |
| } |
| return (response); |
| } |
| |
| /* |
| * Process an ingnore wide residue message. |
| */ |
| static void |
| ahd_handle_ign_wide_residue(struct ahd_softc *ahd, struct ahd_devinfo *devinfo) |
| { |
| u_int scb_index; |
| struct scb *scb; |
| |
| scb_index = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scb_index); |
| /* |
| * XXX Actually check data direction in the sequencer? |
| * Perhaps add datadir to some spare bits in the hscb? |
| */ |
| if ((ahd_inb(ahd, SEQ_FLAGS) & DPHASE) == 0 |
| || ahd_get_transfer_dir(scb) != CAM_DIR_IN) { |
| /* |
| * Ignore the message if we haven't |
| * seen an appropriate data phase yet. |
| */ |
| } else { |
| /* |
| * If the residual occurred on the last |
| * transfer and the transfer request was |
| * expected to end on an odd count, do |
| * nothing. Otherwise, subtract a byte |
| * and update the residual count accordingly. |
| */ |
| uint32_t sgptr; |
| |
| sgptr = ahd_inb_scbram(ahd, SCB_RESIDUAL_SGPTR); |
| if ((sgptr & SG_LIST_NULL) != 0 |
| && (ahd_inb_scbram(ahd, SCB_TASK_ATTRIBUTE) |
| & SCB_XFERLEN_ODD) != 0) { |
| /* |
| * If the residual occurred on the last |
| * transfer and the transfer request was |
| * expected to end on an odd count, do |
| * nothing. |
| */ |
| } else { |
| uint32_t data_cnt; |
| uint64_t data_addr; |
| uint32_t sglen; |
| |
| /* Pull in the rest of the sgptr */ |
| sgptr = ahd_inl_scbram(ahd, SCB_RESIDUAL_SGPTR); |
| data_cnt = ahd_inl_scbram(ahd, SCB_RESIDUAL_DATACNT); |
| if ((sgptr & SG_LIST_NULL) != 0) { |
| /* |
| * The residual data count is not updated |
| * for the command run to completion case. |
| * Explicitly zero the count. |
| */ |
| data_cnt &= ~AHD_SG_LEN_MASK; |
| } |
| data_addr = ahd_inq(ahd, SHADDR); |
| data_cnt += 1; |
| data_addr -= 1; |
| sgptr &= SG_PTR_MASK; |
| if ((ahd->flags & AHD_64BIT_ADDRESSING) != 0) { |
| struct ahd_dma64_seg *sg; |
| |
| sg = ahd_sg_bus_to_virt(ahd, scb, sgptr); |
| |
| /* |
| * The residual sg ptr points to the next S/G |
| * to load so we must go back one. |
| */ |
| sg--; |
| sglen = ahd_le32toh(sg->len) & AHD_SG_LEN_MASK; |
| if (sg != scb->sg_list |
| && sglen < (data_cnt & AHD_SG_LEN_MASK)) { |
| |
| sg--; |
| sglen = ahd_le32toh(sg->len); |
| /* |
| * Preserve High Address and SG_LIST |
| * bits while setting the count to 1. |
| */ |
| data_cnt = 1|(sglen&(~AHD_SG_LEN_MASK)); |
| data_addr = ahd_le64toh(sg->addr) |
| + (sglen & AHD_SG_LEN_MASK) |
| - 1; |
| |
| /* |
| * Increment sg so it points to the |
| * "next" sg. |
| */ |
| sg++; |
| sgptr = ahd_sg_virt_to_bus(ahd, scb, |
| sg); |
| } |
| } else { |
| struct ahd_dma_seg *sg; |
| |
| sg = ahd_sg_bus_to_virt(ahd, scb, sgptr); |
| |
| /* |
| * The residual sg ptr points to the next S/G |
| * to load so we must go back one. |
| */ |
| sg--; |
| sglen = ahd_le32toh(sg->len) & AHD_SG_LEN_MASK; |
| if (sg != scb->sg_list |
| && sglen < (data_cnt & AHD_SG_LEN_MASK)) { |
| |
| sg--; |
| sglen = ahd_le32toh(sg->len); |
| /* |
| * Preserve High Address and SG_LIST |
| * bits while setting the count to 1. |
| */ |
| data_cnt = 1|(sglen&(~AHD_SG_LEN_MASK)); |
| data_addr = ahd_le32toh(sg->addr) |
| + (sglen & AHD_SG_LEN_MASK) |
| - 1; |
| |
| /* |
| * Increment sg so it points to the |
| * "next" sg. |
| */ |
| sg++; |
| sgptr = ahd_sg_virt_to_bus(ahd, scb, |
| sg); |
| } |
| } |
| /* |
| * Toggle the "oddness" of the transfer length |
| * to handle this mid-transfer ignore wide |
| * residue. This ensures that the oddness is |
| * correct for subsequent data transfers. |
| */ |
| ahd_outb(ahd, SCB_TASK_ATTRIBUTE, |
| ahd_inb_scbram(ahd, SCB_TASK_ATTRIBUTE) |
| ^ SCB_XFERLEN_ODD); |
| |
| ahd_outl(ahd, SCB_RESIDUAL_SGPTR, sgptr); |
| ahd_outl(ahd, SCB_RESIDUAL_DATACNT, data_cnt); |
| /* |
| * The FIFO's pointers will be updated if/when the |
| * sequencer re-enters a data phase. |
| */ |
| } |
| } |
| } |
| |
| |
| /* |
| * Reinitialize the data pointers for the active transfer |
| * based on its current residual. |
| */ |
| static void |
| ahd_reinitialize_dataptrs(struct ahd_softc *ahd) |
| { |
| struct scb *scb; |
| ahd_mode_state saved_modes; |
| u_int scb_index; |
| u_int wait; |
| uint32_t sgptr; |
| uint32_t resid; |
| uint64_t dataptr; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_DFF0_MSK|AHD_MODE_DFF1_MSK, |
| AHD_MODE_DFF0_MSK|AHD_MODE_DFF1_MSK); |
| |
| scb_index = ahd_get_scbptr(ahd); |
| scb = ahd_lookup_scb(ahd, scb_index); |
| |
| /* |
| * Release and reacquire the FIFO so we |
| * have a clean slate. |
| */ |
| ahd_outb(ahd, DFFSXFRCTL, CLRCHN); |
| wait = 1000; |
| while (--wait && !(ahd_inb(ahd, MDFFSTAT) & FIFOFREE)) |
| ahd_delay(100); |
| if (wait == 0) { |
| ahd_print_path(ahd, scb); |
| printk("ahd_reinitialize_dataptrs: Forcing FIFO free.\n"); |
| ahd_outb(ahd, DFFSXFRCTL, RSTCHN|CLRSHCNT); |
| } |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| ahd_outb(ahd, DFFSTAT, |
| ahd_inb(ahd, DFFSTAT) |
| | (saved_modes == 0x11 ? CURRFIFO_1 : CURRFIFO_0)); |
| |
| /* |
| * Determine initial values for data_addr and data_cnt |
| * for resuming the data phase. |
| */ |
| sgptr = ahd_inl_scbram(ahd, SCB_RESIDUAL_SGPTR); |
| sgptr &= SG_PTR_MASK; |
| |
| resid = (ahd_inb_scbram(ahd, SCB_RESIDUAL_DATACNT + 2) << 16) |
| | (ahd_inb_scbram(ahd, SCB_RESIDUAL_DATACNT + 1) << 8) |
| | ahd_inb_scbram(ahd, SCB_RESIDUAL_DATACNT); |
| |
| if ((ahd->flags & AHD_64BIT_ADDRESSING) != 0) { |
| struct ahd_dma64_seg *sg; |
| |
| sg = ahd_sg_bus_to_virt(ahd, scb, sgptr); |
| |
| /* The residual sg_ptr always points to the next sg */ |
| sg--; |
| |
| dataptr = ahd_le64toh(sg->addr) |
| + (ahd_le32toh(sg->len) & AHD_SG_LEN_MASK) |
| - resid; |
| ahd_outl(ahd, HADDR + 4, dataptr >> 32); |
| } else { |
| struct ahd_dma_seg *sg; |
| |
| sg = ahd_sg_bus_to_virt(ahd, scb, sgptr); |
| |
| /* The residual sg_ptr always points to the next sg */ |
| sg--; |
| |
| dataptr = ahd_le32toh(sg->addr) |
| + (ahd_le32toh(sg->len) & AHD_SG_LEN_MASK) |
| - resid; |
| ahd_outb(ahd, HADDR + 4, |
| (ahd_le32toh(sg->len) & ~AHD_SG_LEN_MASK) >> 24); |
| } |
| ahd_outl(ahd, HADDR, dataptr); |
| ahd_outb(ahd, HCNT + 2, resid >> 16); |
| ahd_outb(ahd, HCNT + 1, resid >> 8); |
| ahd_outb(ahd, HCNT, resid); |
| } |
| |
| /* |
| * Handle the effects of issuing a bus device reset message. |
| */ |
| static void |
| ahd_handle_devreset(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| u_int lun, cam_status status, char *message, |
| int verbose_level) |
| { |
| #ifdef AHD_TARGET_MODE |
| struct ahd_tmode_tstate* tstate; |
| #endif |
| int found; |
| |
| found = ahd_abort_scbs(ahd, devinfo->target, devinfo->channel, |
| lun, SCB_LIST_NULL, devinfo->role, |
| status); |
| |
| #ifdef AHD_TARGET_MODE |
| /* |
| * Send an immediate notify ccb to all target mord peripheral |
| * drivers affected by this action. |
| */ |
| tstate = ahd->enabled_targets[devinfo->our_scsiid]; |
| if (tstate != NULL) { |
| u_int cur_lun; |
| u_int max_lun; |
| |
| if (lun != CAM_LUN_WILDCARD) { |
| cur_lun = 0; |
| max_lun = AHD_NUM_LUNS - 1; |
| } else { |
| cur_lun = lun; |
| max_lun = lun; |
| } |
| for (;cur_lun <= max_lun; cur_lun++) { |
| struct ahd_tmode_lstate* lstate; |
| |
| lstate = tstate->enabled_luns[cur_lun]; |
| if (lstate == NULL) |
| continue; |
| |
| ahd_queue_lstate_event(ahd, lstate, devinfo->our_scsiid, |
| MSG_BUS_DEV_RESET, /*arg*/0); |
| ahd_send_lstate_events(ahd, lstate); |
| } |
| } |
| #endif |
| |
| /* |
| * Go back to async/narrow transfers and renegotiate. |
| */ |
| ahd_set_width(ahd, devinfo, MSG_EXT_WDTR_BUS_8_BIT, |
| AHD_TRANS_CUR, /*paused*/TRUE); |
| ahd_set_syncrate(ahd, devinfo, /*period*/0, /*offset*/0, |
| /*ppr_options*/0, AHD_TRANS_CUR, |
| /*paused*/TRUE); |
| |
| if (status != CAM_SEL_TIMEOUT) |
| ahd_send_async(ahd, devinfo->channel, devinfo->target, |
| CAM_LUN_WILDCARD, AC_SENT_BDR); |
| |
| if (message != NULL && bootverbose) |
| printk("%s: %s on %c:%d. %d SCBs aborted\n", ahd_name(ahd), |
| message, devinfo->channel, devinfo->target, found); |
| } |
| |
| #ifdef AHD_TARGET_MODE |
| static void |
| ahd_setup_target_msgin(struct ahd_softc *ahd, struct ahd_devinfo *devinfo, |
| struct scb *scb) |
| { |
| |
| /* |
| * To facilitate adding multiple messages together, |
| * each routine should increment the index and len |
| * variables instead of setting them explicitly. |
| */ |
| ahd->msgout_index = 0; |
| ahd->msgout_len = 0; |
| |
| if (scb != NULL && (scb->flags & SCB_AUTO_NEGOTIATE) != 0) |
| ahd_build_transfer_msg(ahd, devinfo); |
| else |
| panic("ahd_intr: AWAITING target message with no message"); |
| |
| ahd->msgout_index = 0; |
| ahd->msg_type = MSG_TYPE_TARGET_MSGIN; |
| } |
| #endif |
| /**************************** Initialization **********************************/ |
| static u_int |
| ahd_sglist_size(struct ahd_softc *ahd) |
| { |
| bus_size_t list_size; |
| |
| list_size = sizeof(struct ahd_dma_seg) * AHD_NSEG; |
| if ((ahd->flags & AHD_64BIT_ADDRESSING) != 0) |
| list_size = sizeof(struct ahd_dma64_seg) * AHD_NSEG; |
| return (list_size); |
| } |
| |
| /* |
| * Calculate the optimum S/G List allocation size. S/G elements used |
| * for a given transaction must be physically contiguous. Assume the |
| * OS will allocate full pages to us, so it doesn't make sense to request |
| * less than a page. |
| */ |
| static u_int |
| ahd_sglist_allocsize(struct ahd_softc *ahd) |
| { |
| bus_size_t sg_list_increment; |
| bus_size_t sg_list_size; |
| bus_size_t max_list_size; |
| bus_size_t best_list_size; |
| |
| /* Start out with the minimum required for AHD_NSEG. */ |
| sg_list_increment = ahd_sglist_size(ahd); |
| sg_list_size = sg_list_increment; |
| |
| /* Get us as close as possible to a page in size. */ |
| while ((sg_list_size + sg_list_increment) <= PAGE_SIZE) |
| sg_list_size += sg_list_increment; |
| |
| /* |
| * Try to reduce the amount of wastage by allocating |
| * multiple pages. |
| */ |
| best_list_size = sg_list_size; |
| max_list_size = roundup(sg_list_increment, PAGE_SIZE); |
| if (max_list_size < 4 * PAGE_SIZE) |
| max_list_size = 4 * PAGE_SIZE; |
| if (max_list_size > (AHD_SCB_MAX_ALLOC * sg_list_increment)) |
| max_list_size = (AHD_SCB_MAX_ALLOC * sg_list_increment); |
| while ((sg_list_size + sg_list_increment) <= max_list_size |
| && (sg_list_size % PAGE_SIZE) != 0) { |
| bus_size_t new_mod; |
| bus_size_t best_mod; |
| |
| sg_list_size += sg_list_increment; |
| new_mod = sg_list_size % PAGE_SIZE; |
| best_mod = best_list_size % PAGE_SIZE; |
| if (new_mod > best_mod || new_mod == 0) { |
| best_list_size = sg_list_size; |
| } |
| } |
| return (best_list_size); |
| } |
| |
| /* |
| * Allocate a controller structure for a new device |
| * and perform initial initializion. |
| */ |
| struct ahd_softc * |
| ahd_alloc(void *platform_arg, char *name) |
| { |
| struct ahd_softc *ahd; |
| |
| #ifndef __FreeBSD__ |
| ahd = kmalloc(sizeof(*ahd), GFP_ATOMIC); |
| if (!ahd) { |
| printk("aic7xxx: cannot malloc softc!\n"); |
| kfree(name); |
| return NULL; |
| } |
| #else |
| ahd = device_get_softc((device_t)platform_arg); |
| #endif |
| memset(ahd, 0, sizeof(*ahd)); |
| ahd->seep_config = kmalloc(sizeof(*ahd->seep_config), GFP_ATOMIC); |
| if (ahd->seep_config == NULL) { |
| #ifndef __FreeBSD__ |
| kfree(ahd); |
| #endif |
| kfree(name); |
| return (NULL); |
| } |
| LIST_INIT(&ahd->pending_scbs); |
| /* We don't know our unit number until the OSM sets it */ |
| ahd->name = name; |
| ahd->unit = -1; |
| ahd->description = NULL; |
| ahd->bus_description = NULL; |
| ahd->channel = 'A'; |
| ahd->chip = AHD_NONE; |
| ahd->features = AHD_FENONE; |
| ahd->bugs = AHD_BUGNONE; |
| ahd->flags = AHD_SPCHK_ENB_A|AHD_RESET_BUS_A|AHD_TERM_ENB_A |
| | AHD_EXTENDED_TRANS_A|AHD_STPWLEVEL_A; |
| ahd_timer_init(&ahd->reset_timer); |
| ahd_timer_init(&ahd->stat_timer); |
| ahd->int_coalescing_timer = AHD_INT_COALESCING_TIMER_DEFAULT; |
| ahd->int_coalescing_maxcmds = AHD_INT_COALESCING_MAXCMDS_DEFAULT; |
| ahd->int_coalescing_mincmds = AHD_INT_COALESCING_MINCMDS_DEFAULT; |
| ahd->int_coalescing_threshold = AHD_INT_COALESCING_THRESHOLD_DEFAULT; |
| ahd->int_coalescing_stop_threshold = |
| AHD_INT_COALESCING_STOP_THRESHOLD_DEFAULT; |
| |
| if (ahd_platform_alloc(ahd, platform_arg) != 0) { |
| ahd_free(ahd); |
| ahd = NULL; |
| } |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MEMORY) != 0) { |
| printk("%s: scb size = 0x%x, hscb size = 0x%x\n", |
| ahd_name(ahd), (u_int)sizeof(struct scb), |
| (u_int)sizeof(struct hardware_scb)); |
| } |
| #endif |
| return (ahd); |
| } |
| |
| int |
| ahd_softc_init(struct ahd_softc *ahd) |
| { |
| |
| ahd->unpause = 0; |
| ahd->pause = PAUSE; |
| return (0); |
| } |
| |
| void |
| ahd_set_unit(struct ahd_softc *ahd, int unit) |
| { |
| ahd->unit = unit; |
| } |
| |
| void |
| ahd_set_name(struct ahd_softc *ahd, char *name) |
| { |
| if (ahd->name != NULL) |
| kfree(ahd->name); |
| ahd->name = name; |
| } |
| |
| void |
| ahd_free(struct ahd_softc *ahd) |
| { |
| int i; |
| |
| switch (ahd->init_level) { |
| default: |
| case 5: |
| ahd_shutdown(ahd); |
| /* FALLTHROUGH */ |
| case 4: |
| ahd_dmamap_unload(ahd, ahd->shared_data_dmat, |
| ahd->shared_data_map.dmamap); |
| /* FALLTHROUGH */ |
| case 3: |
| ahd_dmamem_free(ahd, ahd->shared_data_dmat, ahd->qoutfifo, |
| ahd->shared_data_map.dmamap); |
| ahd_dmamap_destroy(ahd, ahd->shared_data_dmat, |
| ahd->shared_data_map.dmamap); |
| /* FALLTHROUGH */ |
| case 2: |
| ahd_dma_tag_destroy(ahd, ahd->shared_data_dmat); |
| case 1: |
| #ifndef __linux__ |
| ahd_dma_tag_destroy(ahd, ahd->buffer_dmat); |
| #endif |
| break; |
| case 0: |
| break; |
| } |
| |
| #ifndef __linux__ |
| ahd_dma_tag_destroy(ahd, ahd->parent_dmat); |
| #endif |
| ahd_platform_free(ahd); |
| ahd_fini_scbdata(ahd); |
| for (i = 0; i < AHD_NUM_TARGETS; i++) { |
| struct ahd_tmode_tstate *tstate; |
| |
| tstate = ahd->enabled_targets[i]; |
| if (tstate != NULL) { |
| #ifdef AHD_TARGET_MODE |
| int j; |
| |
| for (j = 0; j < AHD_NUM_LUNS; j++) { |
| struct ahd_tmode_lstate *lstate; |
| |
| lstate = tstate->enabled_luns[j]; |
| if (lstate != NULL) { |
| xpt_free_path(lstate->path); |
| kfree(lstate); |
| } |
| } |
| #endif |
| kfree(tstate); |
| } |
| } |
| #ifdef AHD_TARGET_MODE |
| if (ahd->black_hole != NULL) { |
| xpt_free_path(ahd->black_hole->path); |
| kfree(ahd->black_hole); |
| } |
| #endif |
| if (ahd->name != NULL) |
| kfree(ahd->name); |
| if (ahd->seep_config != NULL) |
| kfree(ahd->seep_config); |
| if (ahd->saved_stack != NULL) |
| kfree(ahd->saved_stack); |
| #ifndef __FreeBSD__ |
| kfree(ahd); |
| #endif |
| return; |
| } |
| |
| static void |
| ahd_shutdown(void *arg) |
| { |
| struct ahd_softc *ahd; |
| |
| ahd = (struct ahd_softc *)arg; |
| |
| /* |
| * Stop periodic timer callbacks. |
| */ |
| ahd_timer_stop(&ahd->reset_timer); |
| ahd_timer_stop(&ahd->stat_timer); |
| |
| /* This will reset most registers to 0, but not all */ |
| ahd_reset(ahd, /*reinit*/FALSE); |
| } |
| |
| /* |
| * Reset the controller and record some information about it |
| * that is only available just after a reset. If "reinit" is |
| * non-zero, this reset occurred after initial configuration |
| * and the caller requests that the chip be fully reinitialized |
| * to a runable state. Chip interrupts are *not* enabled after |
| * a reinitialization. The caller must enable interrupts via |
| * ahd_intr_enable(). |
| */ |
| int |
| ahd_reset(struct ahd_softc *ahd, int reinit) |
| { |
| u_int sxfrctl1; |
| int wait; |
| uint32_t cmd; |
| |
| /* |
| * Preserve the value of the SXFRCTL1 register for all channels. |
| * It contains settings that affect termination and we don't want |
| * to disturb the integrity of the bus. |
| */ |
| ahd_pause(ahd); |
| ahd_update_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| sxfrctl1 = ahd_inb(ahd, SXFRCTL1); |
| |
| cmd = ahd_pci_read_config(ahd->dev_softc, PCIR_COMMAND, /*bytes*/2); |
| if ((ahd->bugs & AHD_PCIX_CHIPRST_BUG) != 0) { |
| uint32_t mod_cmd; |
| |
| /* |
| * A4 Razor #632 |
| * During the assertion of CHIPRST, the chip |
| * does not disable its parity logic prior to |
| * the start of the reset. This may cause a |
| * parity error to be detected and thus a |
| * spurious SERR or PERR assertion. Disable |
| * PERR and SERR responses during the CHIPRST. |
| */ |
| mod_cmd = cmd & ~(PCIM_CMD_PERRESPEN|PCIM_CMD_SERRESPEN); |
| ahd_pci_write_config(ahd->dev_softc, PCIR_COMMAND, |
| mod_cmd, /*bytes*/2); |
| } |
| ahd_outb(ahd, HCNTRL, CHIPRST | ahd->pause); |
| |
| /* |
| * Ensure that the reset has finished. We delay 1000us |
| * prior to reading the register to make sure the chip |
| * has sufficiently completed its reset to handle register |
| * accesses. |
| */ |
| wait = 1000; |
| do { |
| ahd_delay(1000); |
| } while (--wait && !(ahd_inb(ahd, HCNTRL) & CHIPRSTACK)); |
| |
| if (wait == 0) { |
| printk("%s: WARNING - Failed chip reset! " |
| "Trying to initialize anyway.\n", ahd_name(ahd)); |
| } |
| ahd_outb(ahd, HCNTRL, ahd->pause); |
| |
| if ((ahd->bugs & AHD_PCIX_CHIPRST_BUG) != 0) { |
| /* |
| * Clear any latched PCI error status and restore |
| * previous SERR and PERR response enables. |
| */ |
| ahd_pci_write_config(ahd->dev_softc, PCIR_STATUS + 1, |
| 0xFF, /*bytes*/1); |
| ahd_pci_write_config(ahd->dev_softc, PCIR_COMMAND, |
| cmd, /*bytes*/2); |
| } |
| |
| /* |
| * Mode should be SCSI after a chip reset, but lets |
| * set it just to be safe. We touch the MODE_PTR |
| * register directly so as to bypass the lazy update |
| * code in ahd_set_modes(). |
| */ |
| ahd_known_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| ahd_outb(ahd, MODE_PTR, |
| ahd_build_mode_state(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI)); |
| |
| /* |
| * Restore SXFRCTL1. |
| * |
| * We must always initialize STPWEN to 1 before we |
| * restore the saved values. STPWEN is initialized |
| * to a tri-state condition which can only be cleared |
| * by turning it on. |
| */ |
| ahd_outb(ahd, SXFRCTL1, sxfrctl1|STPWEN); |
| ahd_outb(ahd, SXFRCTL1, sxfrctl1); |
| |
| /* Determine chip configuration */ |
| ahd->features &= ~AHD_WIDE; |
| if ((ahd_inb(ahd, SBLKCTL) & SELWIDE) != 0) |
| ahd->features |= AHD_WIDE; |
| |
| /* |
| * If a recovery action has forced a chip reset, |
| * re-initialize the chip to our liking. |
| */ |
| if (reinit != 0) |
| ahd_chip_init(ahd); |
| |
| return (0); |
| } |
| |
| /* |
| * Determine the number of SCBs available on the controller |
| */ |
| static int |
| ahd_probe_scbs(struct ahd_softc *ahd) { |
| int i; |
| |
| AHD_ASSERT_MODES(ahd, ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK), |
| ~(AHD_MODE_UNKNOWN_MSK|AHD_MODE_CFG_MSK)); |
| for (i = 0; i < AHD_SCB_MAX; i++) { |
| int j; |
| |
| ahd_set_scbptr(ahd, i); |
| ahd_outw(ahd, SCB_BASE, i); |
| for (j = 2; j < 64; j++) |
| ahd_outb(ahd, SCB_BASE+j, 0); |
| /* Start out life as unallocated (needing an abort) */ |
| ahd_outb(ahd, SCB_CONTROL, MK_MESSAGE); |
| if (ahd_inw_scbram(ahd, SCB_BASE) != i) |
| break; |
| ahd_set_scbptr(ahd, 0); |
| if (ahd_inw_scbram(ahd, SCB_BASE) != 0) |
| break; |
| } |
| return (i); |
| } |
| |
| static void |
| ahd_dmamap_cb(void *arg, bus_dma_segment_t *segs, int nseg, int error) |
| { |
| dma_addr_t *baddr; |
| |
| baddr = (dma_addr_t *)arg; |
| *baddr = segs->ds_addr; |
| } |
| |
| static void |
| ahd_initialize_hscbs(struct ahd_softc *ahd) |
| { |
| int i; |
| |
| for (i = 0; i < ahd->scb_data.maxhscbs; i++) { |
| ahd_set_scbptr(ahd, i); |
| |
| /* Clear the control byte. */ |
| ahd_outb(ahd, SCB_CONTROL, 0); |
| |
| /* Set the next pointer */ |
| ahd_outw(ahd, SCB_NEXT, SCB_LIST_NULL); |
| } |
| } |
| |
| static int |
| ahd_init_scbdata(struct ahd_softc *ahd) |
| { |
| struct scb_data *scb_data; |
| int i; |
| |
| scb_data = &ahd->scb_data; |
| TAILQ_INIT(&scb_data->free_scbs); |
| for (i = 0; i < AHD_NUM_TARGETS * AHD_NUM_LUNS_NONPKT; i++) |
| LIST_INIT(&scb_data->free_scb_lists[i]); |
| LIST_INIT(&scb_data->any_dev_free_scb_list); |
| SLIST_INIT(&scb_data->hscb_maps); |
| SLIST_INIT(&scb_data->sg_maps); |
| SLIST_INIT(&scb_data->sense_maps); |
| |
| /* Determine the number of hardware SCBs and initialize them */ |
| scb_data->maxhscbs = ahd_probe_scbs(ahd); |
| if (scb_data->maxhscbs == 0) { |
| printk("%s: No SCB space found\n", ahd_name(ahd)); |
| return (ENXIO); |
| } |
| |
| ahd_initialize_hscbs(ahd); |
| |
| /* |
| * Create our DMA tags. These tags define the kinds of device |
| * accessible memory allocations and memory mappings we will |
| * need to perform during normal operation. |
| * |
| * Unless we need to further restrict the allocation, we rely |
| * on the restrictions of the parent dmat, hence the common |
| * use of MAXADDR and MAXSIZE. |
| */ |
| |
| /* DMA tag for our hardware scb structures */ |
| if (ahd_dma_tag_create(ahd, ahd->parent_dmat, /*alignment*/1, |
| /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1, |
| /*lowaddr*/BUS_SPACE_MAXADDR_32BIT, |
| /*highaddr*/BUS_SPACE_MAXADDR, |
| /*filter*/NULL, /*filterarg*/NULL, |
| PAGE_SIZE, /*nsegments*/1, |
| /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT, |
| /*flags*/0, &scb_data->hscb_dmat) != 0) { |
| goto error_exit; |
| } |
| |
| scb_data->init_level++; |
| |
| /* DMA tag for our S/G structures. */ |
| if (ahd_dma_tag_create(ahd, ahd->parent_dmat, /*alignment*/8, |
| /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1, |
| /*lowaddr*/BUS_SPACE_MAXADDR_32BIT, |
| /*highaddr*/BUS_SPACE_MAXADDR, |
| /*filter*/NULL, /*filterarg*/NULL, |
| ahd_sglist_allocsize(ahd), /*nsegments*/1, |
| /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT, |
| /*flags*/0, &scb_data->sg_dmat) != 0) { |
| goto error_exit; |
| } |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MEMORY) != 0) |
| printk("%s: ahd_sglist_allocsize = 0x%x\n", ahd_name(ahd), |
| ahd_sglist_allocsize(ahd)); |
| #endif |
| |
| scb_data->init_level++; |
| |
| /* DMA tag for our sense buffers. We allocate in page sized chunks */ |
| if (ahd_dma_tag_create(ahd, ahd->parent_dmat, /*alignment*/1, |
| /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1, |
| /*lowaddr*/BUS_SPACE_MAXADDR_32BIT, |
| /*highaddr*/BUS_SPACE_MAXADDR, |
| /*filter*/NULL, /*filterarg*/NULL, |
| PAGE_SIZE, /*nsegments*/1, |
| /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT, |
| /*flags*/0, &scb_data->sense_dmat) != 0) { |
| goto error_exit; |
| } |
| |
| scb_data->init_level++; |
| |
| /* Perform initial CCB allocation */ |
| ahd_alloc_scbs(ahd); |
| |
| if (scb_data->numscbs == 0) { |
| printk("%s: ahd_init_scbdata - " |
| "Unable to allocate initial scbs\n", |
| ahd_name(ahd)); |
| goto error_exit; |
| } |
| |
| /* |
| * Note that we were successful |
| */ |
| return (0); |
| |
| error_exit: |
| |
| return (ENOMEM); |
| } |
| |
| static struct scb * |
| ahd_find_scb_by_tag(struct ahd_softc *ahd, u_int tag) |
| { |
| struct scb *scb; |
| |
| /* |
| * Look on the pending list. |
| */ |
| LIST_FOREACH(scb, &ahd->pending_scbs, pending_links) { |
| if (SCB_GET_TAG(scb) == tag) |
| return (scb); |
| } |
| |
| /* |
| * Then on all of the collision free lists. |
| */ |
| TAILQ_FOREACH(scb, &ahd->scb_data.free_scbs, links.tqe) { |
| struct scb *list_scb; |
| |
| list_scb = scb; |
| do { |
| if (SCB_GET_TAG(list_scb) == tag) |
| return (list_scb); |
| list_scb = LIST_NEXT(list_scb, collision_links); |
| } while (list_scb); |
| } |
| |
| /* |
| * And finally on the generic free list. |
| */ |
| LIST_FOREACH(scb, &ahd->scb_data.any_dev_free_scb_list, links.le) { |
| if (SCB_GET_TAG(scb) == tag) |
| return (scb); |
| } |
| |
| return (NULL); |
| } |
| |
| static void |
| ahd_fini_scbdata(struct ahd_softc *ahd) |
| { |
| struct scb_data *scb_data; |
| |
| scb_data = &ahd->scb_data; |
| if (scb_data == NULL) |
| return; |
| |
| switch (scb_data->init_level) { |
| default: |
| case 7: |
| { |
| struct map_node *sns_map; |
| |
| while ((sns_map = SLIST_FIRST(&scb_data->sense_maps)) != NULL) { |
| SLIST_REMOVE_HEAD(&scb_data->sense_maps, links); |
| ahd_dmamap_unload(ahd, scb_data->sense_dmat, |
| sns_map->dmamap); |
| ahd_dmamem_free(ahd, scb_data->sense_dmat, |
| sns_map->vaddr, sns_map->dmamap); |
| kfree(sns_map); |
| } |
| ahd_dma_tag_destroy(ahd, scb_data->sense_dmat); |
| /* FALLTHROUGH */ |
| } |
| case 6: |
| { |
| struct map_node *sg_map; |
| |
| while ((sg_map = SLIST_FIRST(&scb_data->sg_maps)) != NULL) { |
| SLIST_REMOVE_HEAD(&scb_data->sg_maps, links); |
| ahd_dmamap_unload(ahd, scb_data->sg_dmat, |
| sg_map->dmamap); |
| ahd_dmamem_free(ahd, scb_data->sg_dmat, |
| sg_map->vaddr, sg_map->dmamap); |
| kfree(sg_map); |
| } |
| ahd_dma_tag_destroy(ahd, scb_data->sg_dmat); |
| /* FALLTHROUGH */ |
| } |
| case 5: |
| { |
| struct map_node *hscb_map; |
| |
| while ((hscb_map = SLIST_FIRST(&scb_data->hscb_maps)) != NULL) { |
| SLIST_REMOVE_HEAD(&scb_data->hscb_maps, links); |
| ahd_dmamap_unload(ahd, scb_data->hscb_dmat, |
| hscb_map->dmamap); |
| ahd_dmamem_free(ahd, scb_data->hscb_dmat, |
| hscb_map->vaddr, hscb_map->dmamap); |
| kfree(hscb_map); |
| } |
| ahd_dma_tag_destroy(ahd, scb_data->hscb_dmat); |
| /* FALLTHROUGH */ |
| } |
| case 4: |
| case 3: |
| case 2: |
| case 1: |
| case 0: |
| break; |
| } |
| } |
| |
| /* |
| * DSP filter Bypass must be enabled until the first selection |
| * after a change in bus mode (Razor #491 and #493). |
| */ |
| static void |
| ahd_setup_iocell_workaround(struct ahd_softc *ahd) |
| { |
| ahd_mode_state saved_modes; |
| |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_CFG, AHD_MODE_CFG); |
| ahd_outb(ahd, DSPDATACTL, ahd_inb(ahd, DSPDATACTL) |
| | BYPASSENAB | RCVROFFSTDIS | XMITOFFSTDIS); |
| ahd_outb(ahd, SIMODE0, ahd_inb(ahd, SIMODE0) | (ENSELDO|ENSELDI)); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MISC) != 0) |
| printk("%s: Setting up iocell workaround\n", ahd_name(ahd)); |
| #endif |
| ahd_restore_modes(ahd, saved_modes); |
| ahd->flags &= ~AHD_HAD_FIRST_SEL; |
| } |
| |
| static void |
| ahd_iocell_first_selection(struct ahd_softc *ahd) |
| { |
| ahd_mode_state saved_modes; |
| u_int sblkctl; |
| |
| if ((ahd->flags & AHD_HAD_FIRST_SEL) != 0) |
| return; |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| sblkctl = ahd_inb(ahd, SBLKCTL); |
| ahd_set_modes(ahd, AHD_MODE_CFG, AHD_MODE_CFG); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MISC) != 0) |
| printk("%s: iocell first selection\n", ahd_name(ahd)); |
| #endif |
| if ((sblkctl & ENAB40) != 0) { |
| ahd_outb(ahd, DSPDATACTL, |
| ahd_inb(ahd, DSPDATACTL) & ~BYPASSENAB); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MISC) != 0) |
| printk("%s: BYPASS now disabled\n", ahd_name(ahd)); |
| #endif |
| } |
| ahd_outb(ahd, SIMODE0, ahd_inb(ahd, SIMODE0) & ~(ENSELDO|ENSELDI)); |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| ahd_restore_modes(ahd, saved_modes); |
| ahd->flags |= AHD_HAD_FIRST_SEL; |
| } |
| |
| /*************************** SCB Management ***********************************/ |
| static void |
| ahd_add_col_list(struct ahd_softc *ahd, struct scb *scb, u_int col_idx) |
| { |
| struct scb_list *free_list; |
| struct scb_tailq *free_tailq; |
| struct scb *first_scb; |
| |
| scb->flags |= SCB_ON_COL_LIST; |
| AHD_SET_SCB_COL_IDX(scb, col_idx); |
| free_list = &ahd->scb_data.free_scb_lists[col_idx]; |
| free_tailq = &ahd->scb_data.free_scbs; |
| first_scb = LIST_FIRST(free_list); |
| if (first_scb != NULL) { |
| LIST_INSERT_AFTER(first_scb, scb, collision_links); |
| } else { |
| LIST_INSERT_HEAD(free_list, scb, collision_links); |
| TAILQ_INSERT_TAIL(free_tailq, scb, links.tqe); |
| } |
| } |
| |
| static void |
| ahd_rem_col_list(struct ahd_softc *ahd, struct scb *scb) |
| { |
| struct scb_list *free_list; |
| struct scb_tailq *free_tailq; |
| struct scb *first_scb; |
| u_int col_idx; |
| |
| scb->flags &= ~SCB_ON_COL_LIST; |
| col_idx = AHD_GET_SCB_COL_IDX(ahd, scb); |
| free_list = &ahd->scb_data.free_scb_lists[col_idx]; |
| free_tailq = &ahd->scb_data.free_scbs; |
| first_scb = LIST_FIRST(free_list); |
| if (first_scb == scb) { |
| struct scb *next_scb; |
| |
| /* |
| * Maintain order in the collision free |
| * lists for fairness if this device has |
| * other colliding tags active. |
| */ |
| next_scb = LIST_NEXT(scb, collision_links); |
| if (next_scb != NULL) { |
| TAILQ_INSERT_AFTER(free_tailq, scb, |
| next_scb, links.tqe); |
| } |
| TAILQ_REMOVE(free_tailq, scb, links.tqe); |
| } |
| LIST_REMOVE(scb, collision_links); |
| } |
| |
| /* |
| * Get a free scb. If there are none, see if we can allocate a new SCB. |
| */ |
| struct scb * |
| ahd_get_scb(struct ahd_softc *ahd, u_int col_idx) |
| { |
| struct scb *scb; |
| int tries; |
| |
| tries = 0; |
| look_again: |
| TAILQ_FOREACH(scb, &ahd->scb_data.free_scbs, links.tqe) { |
| if (AHD_GET_SCB_COL_IDX(ahd, scb) != col_idx) { |
| ahd_rem_col_list(ahd, scb); |
| goto found; |
| } |
| } |
| if ((scb = LIST_FIRST(&ahd->scb_data.any_dev_free_scb_list)) == NULL) { |
| |
| if (tries++ != 0) |
| return (NULL); |
| ahd_alloc_scbs(ahd); |
| goto look_again; |
| } |
| LIST_REMOVE(scb, links.le); |
| if (col_idx != AHD_NEVER_COL_IDX |
| && (scb->col_scb != NULL) |
| && (scb->col_scb->flags & SCB_ACTIVE) == 0) { |
| LIST_REMOVE(scb->col_scb, links.le); |
| ahd_add_col_list(ahd, scb->col_scb, col_idx); |
| } |
| found: |
| scb->flags |= SCB_ACTIVE; |
| return (scb); |
| } |
| |
| /* |
| * Return an SCB resource to the free list. |
| */ |
| void |
| ahd_free_scb(struct ahd_softc *ahd, struct scb *scb) |
| { |
| /* Clean up for the next user */ |
| scb->flags = SCB_FLAG_NONE; |
| scb->hscb->control = 0; |
| ahd->scb_data.scbindex[SCB_GET_TAG(scb)] = NULL; |
| |
| if (scb->col_scb == NULL) { |
| |
| /* |
| * No collision possible. Just free normally. |
| */ |
| LIST_INSERT_HEAD(&ahd->scb_data.any_dev_free_scb_list, |
| scb, links.le); |
| } else if ((scb->col_scb->flags & SCB_ON_COL_LIST) != 0) { |
| |
| /* |
| * The SCB we might have collided with is on |
| * a free collision list. Put both SCBs on |
| * the generic list. |
| */ |
| ahd_rem_col_list(ahd, scb->col_scb); |
| LIST_INSERT_HEAD(&ahd->scb_data.any_dev_free_scb_list, |
| scb, links.le); |
| LIST_INSERT_HEAD(&ahd->scb_data.any_dev_free_scb_list, |
| scb->col_scb, links.le); |
| } else if ((scb->col_scb->flags |
| & (SCB_PACKETIZED|SCB_ACTIVE)) == SCB_ACTIVE |
| && (scb->col_scb->hscb->control & TAG_ENB) != 0) { |
| |
| /* |
| * The SCB we might collide with on the next allocation |
| * is still active in a non-packetized, tagged, context. |
| * Put us on the SCB collision list. |
| */ |
| ahd_add_col_list(ahd, scb, |
| AHD_GET_SCB_COL_IDX(ahd, scb->col_scb)); |
| } else { |
| /* |
| * The SCB we might collide with on the next allocation |
| * is either active in a packetized context, or free. |
| * Since we can't collide, put this SCB on the generic |
| * free list. |
| */ |
| LIST_INSERT_HEAD(&ahd->scb_data.any_dev_free_scb_list, |
| scb, links.le); |
| } |
| |
| ahd_platform_scb_free(ahd, scb); |
| } |
| |
| static void |
| ahd_alloc_scbs(struct ahd_softc *ahd) |
| { |
| struct scb_data *scb_data; |
| struct scb *next_scb; |
| struct hardware_scb *hscb; |
| struct map_node *hscb_map; |
| struct map_node *sg_map; |
| struct map_node *sense_map; |
| uint8_t *segs; |
| uint8_t *sense_data; |
| dma_addr_t hscb_busaddr; |
| dma_addr_t sg_busaddr; |
| dma_addr_t sense_busaddr; |
| int newcount; |
| int i; |
| |
| scb_data = &ahd->scb_data; |
| if (scb_data->numscbs >= AHD_SCB_MAX_ALLOC) |
| /* Can't allocate any more */ |
| return; |
| |
| if (scb_data->scbs_left != 0) { |
| int offset; |
| |
| offset = (PAGE_SIZE / sizeof(*hscb)) - scb_data->scbs_left; |
| hscb_map = SLIST_FIRST(&scb_data->hscb_maps); |
| hscb = &((struct hardware_scb *)hscb_map->vaddr)[offset]; |
| hscb_busaddr = hscb_map->physaddr + (offset * sizeof(*hscb)); |
| } else { |
| hscb_map = kmalloc(sizeof(*hscb_map), GFP_ATOMIC); |
| |
| if (hscb_map == NULL) |
| return; |
| |
| /* Allocate the next batch of hardware SCBs */ |
| if (ahd_dmamem_alloc(ahd, scb_data->hscb_dmat, |
| (void **)&hscb_map->vaddr, |
| BUS_DMA_NOWAIT, &hscb_map->dmamap) != 0) { |
| kfree(hscb_map); |
| return; |
| } |
| |
| SLIST_INSERT_HEAD(&scb_data->hscb_maps, hscb_map, links); |
| |
| ahd_dmamap_load(ahd, scb_data->hscb_dmat, hscb_map->dmamap, |
| hscb_map->vaddr, PAGE_SIZE, ahd_dmamap_cb, |
| &hscb_map->physaddr, /*flags*/0); |
| |
| hscb = (struct hardware_scb *)hscb_map->vaddr; |
| hscb_busaddr = hscb_map->physaddr; |
| scb_data->scbs_left = PAGE_SIZE / sizeof(*hscb); |
| } |
| |
| if (scb_data->sgs_left != 0) { |
| int offset; |
| |
| offset = ((ahd_sglist_allocsize(ahd) / ahd_sglist_size(ahd)) |
| - scb_data->sgs_left) * ahd_sglist_size(ahd); |
| sg_map = SLIST_FIRST(&scb_data->sg_maps); |
| segs = sg_map->vaddr + offset; |
| sg_busaddr = sg_map->physaddr + offset; |
| } else { |
| sg_map = kmalloc(sizeof(*sg_map), GFP_ATOMIC); |
| |
| if (sg_map == NULL) |
| return; |
| |
| /* Allocate the next batch of S/G lists */ |
| if (ahd_dmamem_alloc(ahd, scb_data->sg_dmat, |
| (void **)&sg_map->vaddr, |
| BUS_DMA_NOWAIT, &sg_map->dmamap) != 0) { |
| kfree(sg_map); |
| return; |
| } |
| |
| SLIST_INSERT_HEAD(&scb_data->sg_maps, sg_map, links); |
| |
| ahd_dmamap_load(ahd, scb_data->sg_dmat, sg_map->dmamap, |
| sg_map->vaddr, ahd_sglist_allocsize(ahd), |
| ahd_dmamap_cb, &sg_map->physaddr, /*flags*/0); |
| |
| segs = sg_map->vaddr; |
| sg_busaddr = sg_map->physaddr; |
| scb_data->sgs_left = |
| ahd_sglist_allocsize(ahd) / ahd_sglist_size(ahd); |
| #ifdef AHD_DEBUG |
| if (ahd_debug & AHD_SHOW_MEMORY) |
| printk("Mapped SG data\n"); |
| #endif |
| } |
| |
| if (scb_data->sense_left != 0) { |
| int offset; |
| |
| offset = PAGE_SIZE - (AHD_SENSE_BUFSIZE * scb_data->sense_left); |
| sense_map = SLIST_FIRST(&scb_data->sense_maps); |
| sense_data = sense_map->vaddr + offset; |
| sense_busaddr = sense_map->physaddr + offset; |
| } else { |
| sense_map = kmalloc(sizeof(*sense_map), GFP_ATOMIC); |
| |
| if (sense_map == NULL) |
| return; |
| |
| /* Allocate the next batch of sense buffers */ |
| if (ahd_dmamem_alloc(ahd, scb_data->sense_dmat, |
| (void **)&sense_map->vaddr, |
| BUS_DMA_NOWAIT, &sense_map->dmamap) != 0) { |
| kfree(sense_map); |
| return; |
| } |
| |
| SLIST_INSERT_HEAD(&scb_data->sense_maps, sense_map, links); |
| |
| ahd_dmamap_load(ahd, scb_data->sense_dmat, sense_map->dmamap, |
| sense_map->vaddr, PAGE_SIZE, ahd_dmamap_cb, |
| &sense_map->physaddr, /*flags*/0); |
| |
| sense_data = sense_map->vaddr; |
| sense_busaddr = sense_map->physaddr; |
| scb_data->sense_left = PAGE_SIZE / AHD_SENSE_BUFSIZE; |
| #ifdef AHD_DEBUG |
| if (ahd_debug & AHD_SHOW_MEMORY) |
| printk("Mapped sense data\n"); |
| #endif |
| } |
| |
| newcount = min(scb_data->sense_left, scb_data->scbs_left); |
| newcount = min(newcount, scb_data->sgs_left); |
| newcount = min(newcount, (AHD_SCB_MAX_ALLOC - scb_data->numscbs)); |
| for (i = 0; i < newcount; i++) { |
| struct scb_platform_data *pdata; |
| u_int col_tag; |
| #ifndef __linux__ |
| int error; |
| #endif |
| |
| next_scb = kmalloc(sizeof(*next_scb), GFP_ATOMIC); |
| if (next_scb == NULL) |
| break; |
| |
| pdata = kmalloc(sizeof(*pdata), GFP_ATOMIC); |
| if (pdata == NULL) { |
| kfree(next_scb); |
| break; |
| } |
| next_scb->platform_data = pdata; |
| next_scb->hscb_map = hscb_map; |
| next_scb->sg_map = sg_map; |
| next_scb->sense_map = sense_map; |
| next_scb->sg_list = segs; |
| next_scb->sense_data = sense_data; |
| next_scb->sense_busaddr = sense_busaddr; |
| memset(hscb, 0, sizeof(*hscb)); |
| next_scb->hscb = hscb; |
| hscb->hscb_busaddr = ahd_htole32(hscb_busaddr); |
| |
| /* |
| * The sequencer always starts with the second entry. |
| * The first entry is embedded in the scb. |
| */ |
| next_scb->sg_list_busaddr = sg_busaddr; |
| if ((ahd->flags & AHD_64BIT_ADDRESSING) != 0) |
| next_scb->sg_list_busaddr |
| += sizeof(struct ahd_dma64_seg); |
| else |
| next_scb->sg_list_busaddr += sizeof(struct ahd_dma_seg); |
| next_scb->ahd_softc = ahd; |
| next_scb->flags = SCB_FLAG_NONE; |
| #ifndef __linux__ |
| error = ahd_dmamap_create(ahd, ahd->buffer_dmat, /*flags*/0, |
| &next_scb->dmamap); |
| if (error != 0) { |
| kfree(next_scb); |
| kfree(pdata); |
| break; |
| } |
| #endif |
| next_scb->hscb->tag = ahd_htole16(scb_data->numscbs); |
| col_tag = scb_data->numscbs ^ 0x100; |
| next_scb->col_scb = ahd_find_scb_by_tag(ahd, col_tag); |
| if (next_scb->col_scb != NULL) |
| next_scb->col_scb->col_scb = next_scb; |
| ahd_free_scb(ahd, next_scb); |
| hscb++; |
| hscb_busaddr += sizeof(*hscb); |
| segs += ahd_sglist_size(ahd); |
| sg_busaddr += ahd_sglist_size(ahd); |
| sense_data += AHD_SENSE_BUFSIZE; |
| sense_busaddr += AHD_SENSE_BUFSIZE; |
| scb_data->numscbs++; |
| scb_data->sense_left--; |
| scb_data->scbs_left--; |
| scb_data->sgs_left--; |
| } |
| } |
| |
| void |
| ahd_controller_info(struct ahd_softc *ahd, char *buf) |
| { |
| const char *speed; |
| const char *type; |
| int len; |
| |
| len = sprintf(buf, "%s: ", ahd_chip_names[ahd->chip & AHD_CHIPID_MASK]); |
| buf += len; |
| |
| speed = "Ultra320 "; |
| if ((ahd->features & AHD_WIDE) != 0) { |
| type = "Wide "; |
| } else { |
| type = "Single "; |
| } |
| len = sprintf(buf, "%s%sChannel %c, SCSI Id=%d, ", |
| speed, type, ahd->channel, ahd->our_id); |
| buf += len; |
| |
| sprintf(buf, "%s, %d SCBs", ahd->bus_description, |
| ahd->scb_data.maxhscbs); |
| } |
| |
| static const char *channel_strings[] = { |
| "Primary Low", |
| "Primary High", |
| "Secondary Low", |
| "Secondary High" |
| }; |
| |
| static const char *termstat_strings[] = { |
| "Terminated Correctly", |
| "Over Terminated", |
| "Under Terminated", |
| "Not Configured" |
| }; |
| |
| /***************************** Timer Facilities *******************************/ |
| #define ahd_timer_init init_timer |
| #define ahd_timer_stop del_timer_sync |
| typedef void ahd_linux_callback_t (u_long); |
| |
| static void |
| ahd_timer_reset(ahd_timer_t *timer, int usec, ahd_callback_t *func, void *arg) |
| { |
| struct ahd_softc *ahd; |
| |
| ahd = (struct ahd_softc *)arg; |
| del_timer(timer); |
| timer->data = (u_long)arg; |
| timer->expires = jiffies + (usec * HZ)/1000000; |
| timer->function = (ahd_linux_callback_t*)func; |
| add_timer(timer); |
| } |
| |
| /* |
| * Start the board, ready for normal operation |
| */ |
| int |
| ahd_init(struct ahd_softc *ahd) |
| { |
| uint8_t *next_vaddr; |
| dma_addr_t next_baddr; |
| size_t driver_data_size; |
| int i; |
| int error; |
| u_int warn_user; |
| uint8_t current_sensing; |
| uint8_t fstat; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| |
| ahd->stack_size = ahd_probe_stack_size(ahd); |
| ahd->saved_stack = kmalloc(ahd->stack_size * sizeof(uint16_t), GFP_ATOMIC); |
| if (ahd->saved_stack == NULL) |
| return (ENOMEM); |
| |
| /* |
| * Verify that the compiler hasn't over-aggressively |
| * padded important structures. |
| */ |
| if (sizeof(struct hardware_scb) != 64) |
| panic("Hardware SCB size is incorrect"); |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_DEBUG_SEQUENCER) != 0) |
| ahd->flags |= AHD_SEQUENCER_DEBUG; |
| #endif |
| |
| /* |
| * Default to allowing initiator operations. |
| */ |
| ahd->flags |= AHD_INITIATORROLE; |
| |
| /* |
| * Only allow target mode features if this unit has them enabled. |
| */ |
| if ((AHD_TMODE_ENABLE & (0x1 << ahd->unit)) == 0) |
| ahd->features &= ~AHD_TARGETMODE; |
| |
| #ifndef __linux__ |
| /* DMA tag for mapping buffers into device visible space. */ |
| if (ahd_dma_tag_create(ahd, ahd->parent_dmat, /*alignment*/1, |
| /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1, |
| /*lowaddr*/ahd->flags & AHD_39BIT_ADDRESSING |
| ? (dma_addr_t)0x7FFFFFFFFFULL |
| : BUS_SPACE_MAXADDR_32BIT, |
| /*highaddr*/BUS_SPACE_MAXADDR, |
| /*filter*/NULL, /*filterarg*/NULL, |
| /*maxsize*/(AHD_NSEG - 1) * PAGE_SIZE, |
| /*nsegments*/AHD_NSEG, |
| /*maxsegsz*/AHD_MAXTRANSFER_SIZE, |
| /*flags*/BUS_DMA_ALLOCNOW, |
| &ahd->buffer_dmat) != 0) { |
| return (ENOMEM); |
| } |
| #endif |
| |
| ahd->init_level++; |
| |
| /* |
| * DMA tag for our command fifos and other data in system memory |
| * the card's sequencer must be able to access. For initiator |
| * roles, we need to allocate space for the qoutfifo. When providing |
| * for the target mode role, we must additionally provide space for |
| * the incoming target command fifo. |
| */ |
| driver_data_size = AHD_SCB_MAX * sizeof(*ahd->qoutfifo) |
| + sizeof(struct hardware_scb); |
| if ((ahd->features & AHD_TARGETMODE) != 0) |
| driver_data_size += AHD_TMODE_CMDS * sizeof(struct target_cmd); |
| if ((ahd->bugs & AHD_PKT_BITBUCKET_BUG) != 0) |
| driver_data_size += PKT_OVERRUN_BUFSIZE; |
| if (ahd_dma_tag_create(ahd, ahd->parent_dmat, /*alignment*/1, |
| /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1, |
| /*lowaddr*/BUS_SPACE_MAXADDR_32BIT, |
| /*highaddr*/BUS_SPACE_MAXADDR, |
| /*filter*/NULL, /*filterarg*/NULL, |
| driver_data_size, |
| /*nsegments*/1, |
| /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT, |
| /*flags*/0, &ahd->shared_data_dmat) != 0) { |
| return (ENOMEM); |
| } |
| |
| ahd->init_level++; |
| |
| /* Allocation of driver data */ |
| if (ahd_dmamem_alloc(ahd, ahd->shared_data_dmat, |
| (void **)&ahd->shared_data_map.vaddr, |
| BUS_DMA_NOWAIT, |
| &ahd->shared_data_map.dmamap) != 0) { |
| return (ENOMEM); |
| } |
| |
| ahd->init_level++; |
| |
| /* And permanently map it in */ |
| ahd_dmamap_load(ahd, ahd->shared_data_dmat, ahd->shared_data_map.dmamap, |
| ahd->shared_data_map.vaddr, driver_data_size, |
| ahd_dmamap_cb, &ahd->shared_data_map.physaddr, |
| /*flags*/0); |
| ahd->qoutfifo = (struct ahd_completion *)ahd->shared_data_map.vaddr; |
| next_vaddr = (uint8_t *)&ahd->qoutfifo[AHD_QOUT_SIZE]; |
| next_baddr = ahd->shared_data_map.physaddr |
| + AHD_QOUT_SIZE*sizeof(struct ahd_completion); |
| if ((ahd->features & AHD_TARGETMODE) != 0) { |
| ahd->targetcmds = (struct target_cmd *)next_vaddr; |
| next_vaddr += AHD_TMODE_CMDS * sizeof(struct target_cmd); |
| next_baddr += AHD_TMODE_CMDS * sizeof(struct target_cmd); |
| } |
| |
| if ((ahd->bugs & AHD_PKT_BITBUCKET_BUG) != 0) { |
| ahd->overrun_buf = next_vaddr; |
| next_vaddr += PKT_OVERRUN_BUFSIZE; |
| next_baddr += PKT_OVERRUN_BUFSIZE; |
| } |
| |
| /* |
| * We need one SCB to serve as the "next SCB". Since the |
| * tag identifier in this SCB will never be used, there is |
| * no point in using a valid HSCB tag from an SCB pulled from |
| * the standard free pool. So, we allocate this "sentinel" |
| * specially from the DMA safe memory chunk used for the QOUTFIFO. |
| */ |
| ahd->next_queued_hscb = (struct hardware_scb *)next_vaddr; |
| ahd->next_queued_hscb_map = &ahd->shared_data_map; |
| ahd->next_queued_hscb->hscb_busaddr = ahd_htole32(next_baddr); |
| |
| ahd->init_level++; |
| |
| /* Allocate SCB data now that buffer_dmat is initialized */ |
| if (ahd_init_scbdata(ahd) != 0) |
| return (ENOMEM); |
| |
| if ((ahd->flags & AHD_INITIATORROLE) == 0) |
| ahd->flags &= ~AHD_RESET_BUS_A; |
| |
| /* |
| * Before committing these settings to the chip, give |
| * the OSM one last chance to modify our configuration. |
| */ |
| ahd_platform_init(ahd); |
| |
| /* Bring up the chip. */ |
| ahd_chip_init(ahd); |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| |
| if ((ahd->flags & AHD_CURRENT_SENSING) == 0) |
| goto init_done; |
| |
| /* |
| * Verify termination based on current draw and |
| * warn user if the bus is over/under terminated. |
| */ |
| error = ahd_write_flexport(ahd, FLXADDR_ROMSTAT_CURSENSECTL, |
| CURSENSE_ENB); |
| if (error != 0) { |
| printk("%s: current sensing timeout 1\n", ahd_name(ahd)); |
| goto init_done; |
| } |
| for (i = 20, fstat = FLX_FSTAT_BUSY; |
| (fstat & FLX_FSTAT_BUSY) != 0 && i; i--) { |
| error = ahd_read_flexport(ahd, FLXADDR_FLEXSTAT, &fstat); |
| if (error != 0) { |
| printk("%s: current sensing timeout 2\n", |
| ahd_name(ahd)); |
| goto init_done; |
| } |
| } |
| if (i == 0) { |
| printk("%s: Timedout during current-sensing test\n", |
| ahd_name(ahd)); |
| goto init_done; |
| } |
| |
| /* Latch Current Sensing status. */ |
| error = ahd_read_flexport(ahd, FLXADDR_CURRENT_STAT, ¤t_sensing); |
| if (error != 0) { |
| printk("%s: current sensing timeout 3\n", ahd_name(ahd)); |
| goto init_done; |
| } |
| |
| /* Diable current sensing. */ |
| ahd_write_flexport(ahd, FLXADDR_ROMSTAT_CURSENSECTL, 0); |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_TERMCTL) != 0) { |
| printk("%s: current_sensing == 0x%x\n", |
| ahd_name(ahd), current_sensing); |
| } |
| #endif |
| warn_user = 0; |
| for (i = 0; i < 4; i++, current_sensing >>= FLX_CSTAT_SHIFT) { |
| u_int term_stat; |
| |
| term_stat = (current_sensing & FLX_CSTAT_MASK); |
| switch (term_stat) { |
| case FLX_CSTAT_OVER: |
| case FLX_CSTAT_UNDER: |
| warn_user++; |
| case FLX_CSTAT_INVALID: |
| case FLX_CSTAT_OKAY: |
| if (warn_user == 0 && bootverbose == 0) |
| break; |
| printk("%s: %s Channel %s\n", ahd_name(ahd), |
| channel_strings[i], termstat_strings[term_stat]); |
| break; |
| } |
| } |
| if (warn_user) { |
| printk("%s: WARNING. Termination is not configured correctly.\n" |
| "%s: WARNING. SCSI bus operations may FAIL.\n", |
| ahd_name(ahd), ahd_name(ahd)); |
| } |
| init_done: |
| ahd_restart(ahd); |
| ahd_timer_reset(&ahd->stat_timer, AHD_STAT_UPDATE_US, |
| ahd_stat_timer, ahd); |
| return (0); |
| } |
| |
| /* |
| * (Re)initialize chip state after a chip reset. |
| */ |
| static void |
| ahd_chip_init(struct ahd_softc *ahd) |
| { |
| uint32_t busaddr; |
| u_int sxfrctl1; |
| u_int scsiseq_template; |
| u_int wait; |
| u_int i; |
| u_int target; |
| |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| /* |
| * Take the LED out of diagnostic mode |
| */ |
| ahd_outb(ahd, SBLKCTL, ahd_inb(ahd, SBLKCTL) & ~(DIAGLEDEN|DIAGLEDON)); |
| |
| /* |
| * Return HS_MAILBOX to its default value. |
| */ |
| ahd->hs_mailbox = 0; |
| ahd_outb(ahd, HS_MAILBOX, 0); |
| |
| /* Set the SCSI Id, SXFRCTL0, SXFRCTL1, and SIMODE1. */ |
| ahd_outb(ahd, IOWNID, ahd->our_id); |
| ahd_outb(ahd, TOWNID, ahd->our_id); |
| sxfrctl1 = (ahd->flags & AHD_TERM_ENB_A) != 0 ? STPWEN : 0; |
| sxfrctl1 |= (ahd->flags & AHD_SPCHK_ENB_A) != 0 ? ENSPCHK : 0; |
| if ((ahd->bugs & AHD_LONG_SETIMO_BUG) |
| && (ahd->seltime != STIMESEL_MIN)) { |
| /* |
| * The selection timer duration is twice as long |
| * as it should be. Halve it by adding "1" to |
| * the user specified setting. |
| */ |
| sxfrctl1 |= ahd->seltime + STIMESEL_BUG_ADJ; |
| } else { |
| sxfrctl1 |= ahd->seltime; |
| } |
| |
| ahd_outb(ahd, SXFRCTL0, DFON); |
| ahd_outb(ahd, SXFRCTL1, sxfrctl1|ahd->seltime|ENSTIMER|ACTNEGEN); |
| ahd_outb(ahd, SIMODE1, ENSELTIMO|ENSCSIRST|ENSCSIPERR); |
| |
| /* |
| * Now that termination is set, wait for up |
| * to 500ms for our transceivers to settle. If |
| * the adapter does not have a cable attached, |
| * the transceivers may never settle, so don't |
| * complain if we fail here. |
| */ |
| for (wait = 10000; |
| (ahd_inb(ahd, SBLKCTL) & (ENAB40|ENAB20)) == 0 && wait; |
| wait--) |
| ahd_delay(100); |
| |
| /* Clear any false bus resets due to the transceivers settling */ |
| ahd_outb(ahd, CLRSINT1, CLRSCSIRSTI); |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| |
| /* Initialize mode specific S/G state. */ |
| for (i = 0; i < 2; i++) { |
| ahd_set_modes(ahd, AHD_MODE_DFF0 + i, AHD_MODE_DFF0 + i); |
| ahd_outb(ahd, LONGJMP_ADDR + 1, INVALID_ADDR); |
| ahd_outb(ahd, SG_STATE, 0); |
| ahd_outb(ahd, CLRSEQINTSRC, 0xFF); |
| ahd_outb(ahd, SEQIMODE, |
| ENSAVEPTRS|ENCFG4DATA|ENCFG4ISTAT |
| |ENCFG4TSTAT|ENCFG4ICMD|ENCFG4TCMD); |
| } |
| |
| ahd_set_modes(ahd, AHD_MODE_CFG, AHD_MODE_CFG); |
| ahd_outb(ahd, DSCOMMAND0, ahd_inb(ahd, DSCOMMAND0)|MPARCKEN|CACHETHEN); |
| ahd_outb(ahd, DFF_THRSH, RD_DFTHRSH_75|WR_DFTHRSH_75); |
| ahd_outb(ahd, SIMODE0, ENIOERR|ENOVERRUN); |
| ahd_outb(ahd, SIMODE3, ENNTRAMPERR|ENOSRAMPERR); |
| if ((ahd->bugs & AHD_BUSFREEREV_BUG) != 0) { |
| ahd_outb(ahd, OPTIONMODE, AUTOACKEN|AUTO_MSGOUT_DE); |
| } else { |
| ahd_outb(ahd, OPTIONMODE, AUTOACKEN|BUSFREEREV|AUTO_MSGOUT_DE); |
| } |
| ahd_outb(ahd, SCSCHKN, CURRFIFODEF|WIDERESEN|SHVALIDSTDIS); |
| if ((ahd->chip & AHD_BUS_MASK) == AHD_PCIX) |
| /* |
| * Do not issue a target abort when a split completion |
| * error occurs. Let our PCIX interrupt handler deal |
| * with it instead. H2A4 Razor #625 |
| */ |
| ahd_outb(ahd, PCIXCTL, ahd_inb(ahd, PCIXCTL) | SPLTSTADIS); |
| |
| if ((ahd->bugs & AHD_LQOOVERRUN_BUG) != 0) |
| ahd_outb(ahd, LQOSCSCTL, LQONOCHKOVER); |
| |
| /* |
| * Tweak IOCELL settings. |
| */ |
| if ((ahd->flags & AHD_HP_BOARD) != 0) { |
| for (i = 0; i < NUMDSPS; i++) { |
| ahd_outb(ahd, DSPSELECT, i); |
| ahd_outb(ahd, WRTBIASCTL, WRTBIASCTL_HP_DEFAULT); |
| } |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MISC) != 0) |
| printk("%s: WRTBIASCTL now 0x%x\n", ahd_name(ahd), |
| WRTBIASCTL_HP_DEFAULT); |
| #endif |
| } |
| ahd_setup_iocell_workaround(ahd); |
| |
| /* |
| * Enable LQI Manager interrupts. |
| */ |
| ahd_outb(ahd, LQIMODE1, ENLQIPHASE_LQ|ENLQIPHASE_NLQ|ENLIQABORT |
| | ENLQICRCI_LQ|ENLQICRCI_NLQ|ENLQIBADLQI |
| | ENLQIOVERI_LQ|ENLQIOVERI_NLQ); |
| ahd_outb(ahd, LQOMODE0, ENLQOATNLQ|ENLQOATNPKT|ENLQOTCRC); |
| /* |
| * We choose to have the sequencer catch LQOPHCHGINPKT errors |
| * manually for the command phase at the start of a packetized |
| * selection case. ENLQOBUSFREE should be made redundant by |
| * the BUSFREE interrupt, but it seems that some LQOBUSFREE |
| * events fail to assert the BUSFREE interrupt so we must |
| * also enable LQOBUSFREE interrupts. |
| */ |
| ahd_outb(ahd, LQOMODE1, ENLQOBUSFREE); |
| |
| /* |
| * Setup sequencer interrupt handlers. |
| */ |
| ahd_outw(ahd, INTVEC1_ADDR, ahd_resolve_seqaddr(ahd, LABEL_seq_isr)); |
| ahd_outw(ahd, INTVEC2_ADDR, ahd_resolve_seqaddr(ahd, LABEL_timer_isr)); |
| |
| /* |
| * Setup SCB Offset registers. |
| */ |
| if ((ahd->bugs & AHD_PKT_LUN_BUG) != 0) { |
| ahd_outb(ahd, LUNPTR, offsetof(struct hardware_scb, |
| pkt_long_lun)); |
| } else { |
| ahd_outb(ahd, LUNPTR, offsetof(struct hardware_scb, lun)); |
| } |
| ahd_outb(ahd, CMDLENPTR, offsetof(struct hardware_scb, cdb_len)); |
| ahd_outb(ahd, ATTRPTR, offsetof(struct hardware_scb, task_attribute)); |
| ahd_outb(ahd, FLAGPTR, offsetof(struct hardware_scb, task_management)); |
| ahd_outb(ahd, CMDPTR, offsetof(struct hardware_scb, |
| shared_data.idata.cdb)); |
| ahd_outb(ahd, QNEXTPTR, |
| offsetof(struct hardware_scb, next_hscb_busaddr)); |
| ahd_outb(ahd, ABRTBITPTR, MK_MESSAGE_BIT_OFFSET); |
| ahd_outb(ahd, ABRTBYTEPTR, offsetof(struct hardware_scb, control)); |
| if ((ahd->bugs & AHD_PKT_LUN_BUG) != 0) { |
| ahd_outb(ahd, LUNLEN, |
| sizeof(ahd->next_queued_hscb->pkt_long_lun) - 1); |
| } else { |
| ahd_outb(ahd, LUNLEN, LUNLEN_SINGLE_LEVEL_LUN); |
| } |
| ahd_outb(ahd, CDBLIMIT, SCB_CDB_LEN_PTR - 1); |
| ahd_outb(ahd, MAXCMD, 0xFF); |
| ahd_outb(ahd, SCBAUTOPTR, |
| AUSCBPTR_EN | offsetof(struct hardware_scb, tag)); |
| |
| /* We haven't been enabled for target mode yet. */ |
| ahd_outb(ahd, MULTARGID, 0); |
| ahd_outb(ahd, MULTARGID + 1, 0); |
| |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| /* Initialize the negotiation table. */ |
| if ((ahd->features & AHD_NEW_IOCELL_OPTS) == 0) { |
| /* |
| * Clear the spare bytes in the neg table to avoid |
| * spurious parity errors. |
| */ |
| for (target = 0; target < AHD_NUM_TARGETS; target++) { |
| ahd_outb(ahd, NEGOADDR, target); |
| ahd_outb(ahd, ANNEXCOL, AHD_ANNEXCOL_PER_DEV0); |
| for (i = 0; i < AHD_NUM_PER_DEV_ANNEXCOLS; i++) |
| ahd_outb(ahd, ANNEXDAT, 0); |
| } |
| } |
| for (target = 0; target < AHD_NUM_TARGETS; target++) { |
| struct ahd_devinfo devinfo; |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_tmode_tstate *tstate; |
| |
| tinfo = ahd_fetch_transinfo(ahd, 'A', ahd->our_id, |
| target, &tstate); |
| ahd_compile_devinfo(&devinfo, ahd->our_id, |
| target, CAM_LUN_WILDCARD, |
| 'A', ROLE_INITIATOR); |
| ahd_update_neg_table(ahd, &devinfo, &tinfo->curr); |
| } |
| |
| ahd_outb(ahd, CLRSINT3, NTRAMPERR|OSRAMPERR); |
| ahd_outb(ahd, CLRINT, CLRSCSIINT); |
| |
| #ifdef NEEDS_MORE_TESTING |
| /* |
| * Always enable abort on incoming L_Qs if this feature is |
| * supported. We use this to catch invalid SCB references. |
| */ |
| if ((ahd->bugs & AHD_ABORT_LQI_BUG) == 0) |
| ahd_outb(ahd, LQCTL1, ABORTPENDING); |
| else |
| #endif |
| ahd_outb(ahd, LQCTL1, 0); |
| |
| /* All of our queues are empty */ |
| ahd->qoutfifonext = 0; |
| ahd->qoutfifonext_valid_tag = QOUTFIFO_ENTRY_VALID; |
| ahd_outb(ahd, QOUTFIFO_ENTRY_VALID_TAG, QOUTFIFO_ENTRY_VALID); |
| for (i = 0; i < AHD_QOUT_SIZE; i++) |
| ahd->qoutfifo[i].valid_tag = 0; |
| ahd_sync_qoutfifo(ahd, BUS_DMASYNC_PREREAD); |
| |
| ahd->qinfifonext = 0; |
| for (i = 0; i < AHD_QIN_SIZE; i++) |
| ahd->qinfifo[i] = SCB_LIST_NULL; |
| |
| if ((ahd->features & AHD_TARGETMODE) != 0) { |
| /* All target command blocks start out invalid. */ |
| for (i = 0; i < AHD_TMODE_CMDS; i++) |
| ahd->targetcmds[i].cmd_valid = 0; |
| ahd_sync_tqinfifo(ahd, BUS_DMASYNC_PREREAD); |
| ahd->tqinfifonext = 1; |
| ahd_outb(ahd, KERNEL_TQINPOS, ahd->tqinfifonext - 1); |
| ahd_outb(ahd, TQINPOS, ahd->tqinfifonext); |
| } |
| |
| /* Initialize Scratch Ram. */ |
| ahd_outb(ahd, SEQ_FLAGS, 0); |
| ahd_outb(ahd, SEQ_FLAGS2, 0); |
| |
| /* We don't have any waiting selections */ |
| ahd_outw(ahd, WAITING_TID_HEAD, SCB_LIST_NULL); |
| ahd_outw(ahd, WAITING_TID_TAIL, SCB_LIST_NULL); |
| ahd_outw(ahd, MK_MESSAGE_SCB, SCB_LIST_NULL); |
| ahd_outw(ahd, MK_MESSAGE_SCSIID, 0xFF); |
| for (i = 0; i < AHD_NUM_TARGETS; i++) |
| ahd_outw(ahd, WAITING_SCB_TAILS + (2 * i), SCB_LIST_NULL); |
| |
| /* |
| * Nobody is waiting to be DMAed into the QOUTFIFO. |
| */ |
| ahd_outw(ahd, COMPLETE_SCB_HEAD, SCB_LIST_NULL); |
| ahd_outw(ahd, COMPLETE_SCB_DMAINPROG_HEAD, SCB_LIST_NULL); |
| ahd_outw(ahd, COMPLETE_DMA_SCB_HEAD, SCB_LIST_NULL); |
| ahd_outw(ahd, COMPLETE_DMA_SCB_TAIL, SCB_LIST_NULL); |
| ahd_outw(ahd, COMPLETE_ON_QFREEZE_HEAD, SCB_LIST_NULL); |
| |
| /* |
| * The Freeze Count is 0. |
| */ |
| ahd->qfreeze_cnt = 0; |
| ahd_outw(ahd, QFREEZE_COUNT, 0); |
| ahd_outw(ahd, KERNEL_QFREEZE_COUNT, 0); |
| |
| /* |
| * Tell the sequencer where it can find our arrays in memory. |
| */ |
| busaddr = ahd->shared_data_map.physaddr; |
| ahd_outl(ahd, SHARED_DATA_ADDR, busaddr); |
| ahd_outl(ahd, QOUTFIFO_NEXT_ADDR, busaddr); |
| |
| /* |
| * Setup the allowed SCSI Sequences based on operational mode. |
| * If we are a target, we'll enable select in operations once |
| * we've had a lun enabled. |
| */ |
| scsiseq_template = ENAUTOATNP; |
| if ((ahd->flags & AHD_INITIATORROLE) != 0) |
| scsiseq_template |= ENRSELI; |
| ahd_outb(ahd, SCSISEQ_TEMPLATE, scsiseq_template); |
| |
| /* There are no busy SCBs yet. */ |
| for (target = 0; target < AHD_NUM_TARGETS; target++) { |
| int lun; |
| |
| for (lun = 0; lun < AHD_NUM_LUNS_NONPKT; lun++) |
| ahd_unbusy_tcl(ahd, BUILD_TCL_RAW(target, 'A', lun)); |
| } |
| |
| /* |
| * Initialize the group code to command length table. |
| * Vendor Unique codes are set to 0 so we only capture |
| * the first byte of the cdb. These can be overridden |
| * when target mode is enabled. |
| */ |
| ahd_outb(ahd, CMDSIZE_TABLE, 5); |
| ahd_outb(ahd, CMDSIZE_TABLE + 1, 9); |
| ahd_outb(ahd, CMDSIZE_TABLE + 2, 9); |
| ahd_outb(ahd, CMDSIZE_TABLE + 3, 0); |
| ahd_outb(ahd, CMDSIZE_TABLE + 4, 15); |
| ahd_outb(ahd, CMDSIZE_TABLE + 5, 11); |
| ahd_outb(ahd, CMDSIZE_TABLE + 6, 0); |
| ahd_outb(ahd, CMDSIZE_TABLE + 7, 0); |
| |
| /* Tell the sequencer of our initial queue positions */ |
| ahd_set_modes(ahd, AHD_MODE_CCHAN, AHD_MODE_CCHAN); |
| ahd_outb(ahd, QOFF_CTLSTA, SCB_QSIZE_512); |
| ahd->qinfifonext = 0; |
| ahd_set_hnscb_qoff(ahd, ahd->qinfifonext); |
| ahd_set_hescb_qoff(ahd, 0); |
| ahd_set_snscb_qoff(ahd, 0); |
| ahd_set_sescb_qoff(ahd, 0); |
| ahd_set_sdscb_qoff(ahd, 0); |
| |
| /* |
| * Tell the sequencer which SCB will be the next one it receives. |
| */ |
| busaddr = ahd_le32toh(ahd->next_queued_hscb->hscb_busaddr); |
| ahd_outl(ahd, NEXT_QUEUED_SCB_ADDR, busaddr); |
| |
| /* |
| * Default to coalescing disabled. |
| */ |
| ahd_outw(ahd, INT_COALESCING_CMDCOUNT, 0); |
| ahd_outw(ahd, CMDS_PENDING, 0); |
| ahd_update_coalescing_values(ahd, ahd->int_coalescing_timer, |
| ahd->int_coalescing_maxcmds, |
| ahd->int_coalescing_mincmds); |
| ahd_enable_coalescing(ahd, FALSE); |
| |
| ahd_loadseq(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| |
| if (ahd->features & AHD_AIC79XXB_SLOWCRC) { |
| u_int negodat3 = ahd_inb(ahd, NEGCONOPTS); |
| |
| negodat3 |= ENSLOWCRC; |
| ahd_outb(ahd, NEGCONOPTS, negodat3); |
| negodat3 = ahd_inb(ahd, NEGCONOPTS); |
| if (!(negodat3 & ENSLOWCRC)) |
| printk("aic79xx: failed to set the SLOWCRC bit\n"); |
| else |
| printk("aic79xx: SLOWCRC bit set\n"); |
| } |
| } |
| |
| /* |
| * Setup default device and controller settings. |
| * This should only be called if our probe has |
| * determined that no configuration data is available. |
| */ |
| int |
| ahd_default_config(struct ahd_softc *ahd) |
| { |
| int targ; |
| |
| ahd->our_id = 7; |
| |
| /* |
| * Allocate a tstate to house information for our |
| * initiator presence on the bus as well as the user |
| * data for any target mode initiator. |
| */ |
| if (ahd_alloc_tstate(ahd, ahd->our_id, 'A') == NULL) { |
| printk("%s: unable to allocate ahd_tmode_tstate. " |
| "Failing attach\n", ahd_name(ahd)); |
| return (ENOMEM); |
| } |
| |
| for (targ = 0; targ < AHD_NUM_TARGETS; targ++) { |
| struct ahd_devinfo devinfo; |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_tmode_tstate *tstate; |
| uint16_t target_mask; |
| |
| tinfo = ahd_fetch_transinfo(ahd, 'A', ahd->our_id, |
| targ, &tstate); |
| /* |
| * We support SPC2 and SPI4. |
| */ |
| tinfo->user.protocol_version = 4; |
| tinfo->user.transport_version = 4; |
| |
| target_mask = 0x01 << targ; |
| ahd->user_discenable |= target_mask; |
| tstate->discenable |= target_mask; |
| ahd->user_tagenable |= target_mask; |
| #ifdef AHD_FORCE_160 |
| tinfo->user.period = AHD_SYNCRATE_DT; |
| #else |
| tinfo->user.period = AHD_SYNCRATE_160; |
| #endif |
| tinfo->user.offset = MAX_OFFSET; |
| tinfo->user.ppr_options = MSG_EXT_PPR_RD_STRM |
| | MSG_EXT_PPR_WR_FLOW |
| | MSG_EXT_PPR_HOLD_MCS |
| | MSG_EXT_PPR_IU_REQ |
| | MSG_EXT_PPR_QAS_REQ |
| | MSG_EXT_PPR_DT_REQ; |
| if ((ahd->features & AHD_RTI) != 0) |
| tinfo->user.ppr_options |= MSG_EXT_PPR_RTI; |
| |
| tinfo->user.width = MSG_EXT_WDTR_BUS_16_BIT; |
| |
| /* |
| * Start out Async/Narrow/Untagged and with |
| * conservative protocol support. |
| */ |
| tinfo->goal.protocol_version = 2; |
| tinfo->goal.transport_version = 2; |
| tinfo->curr.protocol_version = 2; |
| tinfo->curr.transport_version = 2; |
| ahd_compile_devinfo(&devinfo, ahd->our_id, |
| targ, CAM_LUN_WILDCARD, |
| 'A', ROLE_INITIATOR); |
| tstate->tagenable &= ~target_mask; |
| ahd_set_width(ahd, &devinfo, MSG_EXT_WDTR_BUS_8_BIT, |
| AHD_TRANS_CUR|AHD_TRANS_GOAL, /*paused*/TRUE); |
| ahd_set_syncrate(ahd, &devinfo, /*period*/0, /*offset*/0, |
| /*ppr_options*/0, AHD_TRANS_CUR|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| } |
| return (0); |
| } |
| |
| /* |
| * Parse device configuration information. |
| */ |
| int |
| ahd_parse_cfgdata(struct ahd_softc *ahd, struct seeprom_config *sc) |
| { |
| int targ; |
| int max_targ; |
| |
| max_targ = sc->max_targets & CFMAXTARG; |
| ahd->our_id = sc->brtime_id & CFSCSIID; |
| |
| /* |
| * Allocate a tstate to house information for our |
| * initiator presence on the bus as well as the user |
| * data for any target mode initiator. |
| */ |
| if (ahd_alloc_tstate(ahd, ahd->our_id, 'A') == NULL) { |
| printk("%s: unable to allocate ahd_tmode_tstate. " |
| "Failing attach\n", ahd_name(ahd)); |
| return (ENOMEM); |
| } |
| |
| for (targ = 0; targ < max_targ; targ++) { |
| struct ahd_devinfo devinfo; |
| struct ahd_initiator_tinfo *tinfo; |
| struct ahd_transinfo *user_tinfo; |
| struct ahd_tmode_tstate *tstate; |
| uint16_t target_mask; |
| |
| tinfo = ahd_fetch_transinfo(ahd, 'A', ahd->our_id, |
| targ, &tstate); |
| user_tinfo = &tinfo->user; |
| |
| /* |
| * We support SPC2 and SPI4. |
| */ |
| tinfo->user.protocol_version = 4; |
| tinfo->user.transport_version = 4; |
| |
| target_mask = 0x01 << targ; |
| ahd->user_discenable &= ~target_mask; |
| tstate->discenable &= ~target_mask; |
| ahd->user_tagenable &= ~target_mask; |
| if (sc->device_flags[targ] & CFDISC) { |
| tstate->discenable |= target_mask; |
| ahd->user_discenable |= target_mask; |
| ahd->user_tagenable |= target_mask; |
| } else { |
| /* |
| * Cannot be packetized without disconnection. |
| */ |
| sc->device_flags[targ] &= ~CFPACKETIZED; |
| } |
| |
| user_tinfo->ppr_options = 0; |
| user_tinfo->period = (sc->device_flags[targ] & CFXFER); |
| if (user_tinfo->period < CFXFER_ASYNC) { |
| if (user_tinfo->period <= AHD_PERIOD_10MHz) |
| user_tinfo->ppr_options |= MSG_EXT_PPR_DT_REQ; |
| user_tinfo->offset = MAX_OFFSET; |
| } else { |
| user_tinfo->offset = 0; |
| user_tinfo->period = AHD_ASYNC_XFER_PERIOD; |
| } |
| #ifdef AHD_FORCE_160 |
| if (user_tinfo->period <= AHD_SYNCRATE_160) |
| user_tinfo->period = AHD_SYNCRATE_DT; |
| #endif |
| |
| if ((sc->device_flags[targ] & CFPACKETIZED) != 0) { |
| user_tinfo->ppr_options |= MSG_EXT_PPR_RD_STRM |
| | MSG_EXT_PPR_WR_FLOW |
| | MSG_EXT_PPR_HOLD_MCS |
| | MSG_EXT_PPR_IU_REQ; |
| if ((ahd->features & AHD_RTI) != 0) |
| user_tinfo->ppr_options |= MSG_EXT_PPR_RTI; |
| } |
| |
| if ((sc->device_flags[targ] & CFQAS) != 0) |
| user_tinfo->ppr_options |= MSG_EXT_PPR_QAS_REQ; |
| |
| if ((sc->device_flags[targ] & CFWIDEB) != 0) |
| user_tinfo->width = MSG_EXT_WDTR_BUS_16_BIT; |
| else |
| user_tinfo->width = MSG_EXT_WDTR_BUS_8_BIT; |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MISC) != 0) |
| printk("(%d): %x:%x:%x:%x\n", targ, user_tinfo->width, |
| user_tinfo->period, user_tinfo->offset, |
| user_tinfo->ppr_options); |
| #endif |
| /* |
| * Start out Async/Narrow/Untagged and with |
| * conservative protocol support. |
| */ |
| tstate->tagenable &= ~target_mask; |
| tinfo->goal.protocol_version = 2; |
| tinfo->goal.transport_version = 2; |
| tinfo->curr.protocol_version = 2; |
| tinfo->curr.transport_version = 2; |
| ahd_compile_devinfo(&devinfo, ahd->our_id, |
| targ, CAM_LUN_WILDCARD, |
| 'A', ROLE_INITIATOR); |
| ahd_set_width(ahd, &devinfo, MSG_EXT_WDTR_BUS_8_BIT, |
| AHD_TRANS_CUR|AHD_TRANS_GOAL, /*paused*/TRUE); |
| ahd_set_syncrate(ahd, &devinfo, /*period*/0, /*offset*/0, |
| /*ppr_options*/0, AHD_TRANS_CUR|AHD_TRANS_GOAL, |
| /*paused*/TRUE); |
| } |
| |
| ahd->flags &= ~AHD_SPCHK_ENB_A; |
| if (sc->bios_control & CFSPARITY) |
| ahd->flags |= AHD_SPCHK_ENB_A; |
| |
| ahd->flags &= ~AHD_RESET_BUS_A; |
| if (sc->bios_control & CFRESETB) |
| ahd->flags |= AHD_RESET_BUS_A; |
| |
| ahd->flags &= ~AHD_EXTENDED_TRANS_A; |
| if (sc->bios_control & CFEXTEND) |
| ahd->flags |= AHD_EXTENDED_TRANS_A; |
| |
| ahd->flags &= ~AHD_BIOS_ENABLED; |
| if ((sc->bios_control & CFBIOSSTATE) == CFBS_ENABLED) |
| ahd->flags |= AHD_BIOS_ENABLED; |
| |
| ahd->flags &= ~AHD_STPWLEVEL_A; |
| if ((sc->adapter_control & CFSTPWLEVEL) != 0) |
| ahd->flags |= AHD_STPWLEVEL_A; |
| |
| return (0); |
| } |
| |
| /* |
| * Parse device configuration information. |
| */ |
| int |
| ahd_parse_vpddata(struct ahd_softc *ahd, struct vpd_config *vpd) |
| { |
| int error; |
| |
| error = ahd_verify_vpd_cksum(vpd); |
| if (error == 0) |
| return (EINVAL); |
| if ((vpd->bios_flags & VPDBOOTHOST) != 0) |
| ahd->flags |= AHD_BOOT_CHANNEL; |
| return (0); |
| } |
| |
| void |
| ahd_intr_enable(struct ahd_softc *ahd, int enable) |
| { |
| u_int hcntrl; |
| |
| hcntrl = ahd_inb(ahd, HCNTRL); |
| hcntrl &= ~INTEN; |
| ahd->pause &= ~INTEN; |
| ahd->unpause &= ~INTEN; |
| if (enable) { |
| hcntrl |= INTEN; |
| ahd->pause |= INTEN; |
| ahd->unpause |= INTEN; |
| } |
| ahd_outb(ahd, HCNTRL, hcntrl); |
| } |
| |
| static void |
| ahd_update_coalescing_values(struct ahd_softc *ahd, u_int timer, u_int maxcmds, |
| u_int mincmds) |
| { |
| if (timer > AHD_TIMER_MAX_US) |
| timer = AHD_TIMER_MAX_US; |
| ahd->int_coalescing_timer = timer; |
| |
| if (maxcmds > AHD_INT_COALESCING_MAXCMDS_MAX) |
| maxcmds = AHD_INT_COALESCING_MAXCMDS_MAX; |
| if (mincmds > AHD_INT_COALESCING_MINCMDS_MAX) |
| mincmds = AHD_INT_COALESCING_MINCMDS_MAX; |
| ahd->int_coalescing_maxcmds = maxcmds; |
| ahd_outw(ahd, INT_COALESCING_TIMER, timer / AHD_TIMER_US_PER_TICK); |
| ahd_outb(ahd, INT_COALESCING_MAXCMDS, -maxcmds); |
| ahd_outb(ahd, INT_COALESCING_MINCMDS, -mincmds); |
| } |
| |
| static void |
| ahd_enable_coalescing(struct ahd_softc *ahd, int enable) |
| { |
| |
| ahd->hs_mailbox &= ~ENINT_COALESCE; |
| if (enable) |
| ahd->hs_mailbox |= ENINT_COALESCE; |
| ahd_outb(ahd, HS_MAILBOX, ahd->hs_mailbox); |
| ahd_flush_device_writes(ahd); |
| ahd_run_qoutfifo(ahd); |
| } |
| |
| /* |
| * Ensure that the card is paused in a location |
| * outside of all critical sections and that all |
| * pending work is completed prior to returning. |
| * This routine should only be called from outside |
| * an interrupt context. |
| */ |
| void |
| ahd_pause_and_flushwork(struct ahd_softc *ahd) |
| { |
| u_int intstat; |
| u_int maxloops; |
| |
| maxloops = 1000; |
| ahd->flags |= AHD_ALL_INTERRUPTS; |
| ahd_pause(ahd); |
| /* |
| * Freeze the outgoing selections. We do this only |
| * until we are safely paused without further selections |
| * pending. |
| */ |
| ahd->qfreeze_cnt--; |
| ahd_outw(ahd, KERNEL_QFREEZE_COUNT, ahd->qfreeze_cnt); |
| ahd_outb(ahd, SEQ_FLAGS2, ahd_inb(ahd, SEQ_FLAGS2) | SELECTOUT_QFROZEN); |
| do { |
| |
| ahd_unpause(ahd); |
| /* |
| * Give the sequencer some time to service |
| * any active selections. |
| */ |
| ahd_delay(500); |
| |
| ahd_intr(ahd); |
| ahd_pause(ahd); |
| intstat = ahd_inb(ahd, INTSTAT); |
| if ((intstat & INT_PEND) == 0) { |
| ahd_clear_critical_section(ahd); |
| intstat = ahd_inb(ahd, INTSTAT); |
| } |
| } while (--maxloops |
| && (intstat != 0xFF || (ahd->features & AHD_REMOVABLE) == 0) |
| && ((intstat & INT_PEND) != 0 |
| || (ahd_inb(ahd, SCSISEQ0) & ENSELO) != 0 |
| || (ahd_inb(ahd, SSTAT0) & (SELDO|SELINGO)) != 0)); |
| |
| if (maxloops == 0) { |
| printk("Infinite interrupt loop, INTSTAT = %x", |
| ahd_inb(ahd, INTSTAT)); |
| } |
| ahd->qfreeze_cnt++; |
| ahd_outw(ahd, KERNEL_QFREEZE_COUNT, ahd->qfreeze_cnt); |
| |
| ahd_flush_qoutfifo(ahd); |
| |
| ahd->flags &= ~AHD_ALL_INTERRUPTS; |
| } |
| |
| #ifdef CONFIG_PM |
| int |
| ahd_suspend(struct ahd_softc *ahd) |
| { |
| |
| ahd_pause_and_flushwork(ahd); |
| |
| if (LIST_FIRST(&ahd->pending_scbs) != NULL) { |
| ahd_unpause(ahd); |
| return (EBUSY); |
| } |
| ahd_shutdown(ahd); |
| return (0); |
| } |
| |
| void |
| ahd_resume(struct ahd_softc *ahd) |
| { |
| |
| ahd_reset(ahd, /*reinit*/TRUE); |
| ahd_intr_enable(ahd, TRUE); |
| ahd_restart(ahd); |
| } |
| #endif |
| |
| /************************** Busy Target Table *********************************/ |
| /* |
| * Set SCBPTR to the SCB that contains the busy |
| * table entry for TCL. Return the offset into |
| * the SCB that contains the entry for TCL. |
| * saved_scbid is dereferenced and set to the |
| * scbid that should be restored once manipualtion |
| * of the TCL entry is complete. |
| */ |
| static inline u_int |
| ahd_index_busy_tcl(struct ahd_softc *ahd, u_int *saved_scbid, u_int tcl) |
| { |
| /* |
| * Index to the SCB that contains the busy entry. |
| */ |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| *saved_scbid = ahd_get_scbptr(ahd); |
| ahd_set_scbptr(ahd, TCL_LUN(tcl) |
| | ((TCL_TARGET_OFFSET(tcl) & 0xC) << 4)); |
| |
| /* |
| * And now calculate the SCB offset to the entry. |
| * Each entry is 2 bytes wide, hence the |
| * multiplication by 2. |
| */ |
| return (((TCL_TARGET_OFFSET(tcl) & 0x3) << 1) + SCB_DISCONNECTED_LISTS); |
| } |
| |
| /* |
| * Return the untagged transaction id for a given target/channel lun. |
| */ |
| static u_int |
| ahd_find_busy_tcl(struct ahd_softc *ahd, u_int tcl) |
| { |
| u_int scbid; |
| u_int scb_offset; |
| u_int saved_scbptr; |
| |
| scb_offset = ahd_index_busy_tcl(ahd, &saved_scbptr, tcl); |
| scbid = ahd_inw_scbram(ahd, scb_offset); |
| ahd_set_scbptr(ahd, saved_scbptr); |
| return (scbid); |
| } |
| |
| static void |
| ahd_busy_tcl(struct ahd_softc *ahd, u_int tcl, u_int scbid) |
| { |
| u_int scb_offset; |
| u_int saved_scbptr; |
| |
| scb_offset = ahd_index_busy_tcl(ahd, &saved_scbptr, tcl); |
| ahd_outw(ahd, scb_offset, scbid); |
| ahd_set_scbptr(ahd, saved_scbptr); |
| } |
| |
| /************************** SCB and SCB queue management **********************/ |
| static int |
| ahd_match_scb(struct ahd_softc *ahd, struct scb *scb, int target, |
| char channel, int lun, u_int tag, role_t role) |
| { |
| int targ = SCB_GET_TARGET(ahd, scb); |
| char chan = SCB_GET_CHANNEL(ahd, scb); |
| int slun = SCB_GET_LUN(scb); |
| int match; |
| |
| match = ((chan == channel) || (channel == ALL_CHANNELS)); |
| if (match != 0) |
| match = ((targ == target) || (target == CAM_TARGET_WILDCARD)); |
| if (match != 0) |
| match = ((lun == slun) || (lun == CAM_LUN_WILDCARD)); |
| if (match != 0) { |
| #ifdef AHD_TARGET_MODE |
| int group; |
| |
| group = XPT_FC_GROUP(scb->io_ctx->ccb_h.func_code); |
| if (role == ROLE_INITIATOR) { |
| match = (group != XPT_FC_GROUP_TMODE) |
| && ((tag == SCB_GET_TAG(scb)) |
| || (tag == SCB_LIST_NULL)); |
| } else if (role == ROLE_TARGET) { |
| match = (group == XPT_FC_GROUP_TMODE) |
| && ((tag == scb->io_ctx->csio.tag_id) |
| || (tag == SCB_LIST_NULL)); |
| } |
| #else /* !AHD_TARGET_MODE */ |
| match = ((tag == SCB_GET_TAG(scb)) || (tag == SCB_LIST_NULL)); |
| #endif /* AHD_TARGET_MODE */ |
| } |
| |
| return match; |
| } |
| |
| static void |
| ahd_freeze_devq(struct ahd_softc *ahd, struct scb *scb) |
| { |
| int target; |
| char channel; |
| int lun; |
| |
| target = SCB_GET_TARGET(ahd, scb); |
| lun = SCB_GET_LUN(scb); |
| channel = SCB_GET_CHANNEL(ahd, scb); |
| |
| ahd_search_qinfifo(ahd, target, channel, lun, |
| /*tag*/SCB_LIST_NULL, ROLE_UNKNOWN, |
| CAM_REQUEUE_REQ, SEARCH_COMPLETE); |
| |
| ahd_platform_freeze_devq(ahd, scb); |
| } |
| |
| void |
| ahd_qinfifo_requeue_tail(struct ahd_softc *ahd, struct scb *scb) |
| { |
| struct scb *prev_scb; |
| ahd_mode_state saved_modes; |
| |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_CCHAN, AHD_MODE_CCHAN); |
| prev_scb = NULL; |
| if (ahd_qinfifo_count(ahd) != 0) { |
| u_int prev_tag; |
| u_int prev_pos; |
| |
| prev_pos = AHD_QIN_WRAP(ahd->qinfifonext - 1); |
| prev_tag = ahd->qinfifo[prev_pos]; |
| prev_scb = ahd_lookup_scb(ahd, prev_tag); |
| } |
| ahd_qinfifo_requeue(ahd, prev_scb, scb); |
| ahd_set_hnscb_qoff(ahd, ahd->qinfifonext); |
| ahd_restore_modes(ahd, saved_modes); |
| } |
| |
| static void |
| ahd_qinfifo_requeue(struct ahd_softc *ahd, struct scb *prev_scb, |
| struct scb *scb) |
| { |
| if (prev_scb == NULL) { |
| uint32_t busaddr; |
| |
| busaddr = ahd_le32toh(scb->hscb->hscb_busaddr); |
| ahd_outl(ahd, NEXT_QUEUED_SCB_ADDR, busaddr); |
| } else { |
| prev_scb->hscb->next_hscb_busaddr = scb->hscb->hscb_busaddr; |
| ahd_sync_scb(ahd, prev_scb, |
| BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); |
| } |
| ahd->qinfifo[AHD_QIN_WRAP(ahd->qinfifonext)] = SCB_GET_TAG(scb); |
| ahd->qinfifonext++; |
| scb->hscb->next_hscb_busaddr = ahd->next_queued_hscb->hscb_busaddr; |
| ahd_sync_scb(ahd, scb, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); |
| } |
| |
| static int |
| ahd_qinfifo_count(struct ahd_softc *ahd) |
| { |
| u_int qinpos; |
| u_int wrap_qinpos; |
| u_int wrap_qinfifonext; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_CCHAN_MSK, AHD_MODE_CCHAN_MSK); |
| qinpos = ahd_get_snscb_qoff(ahd); |
| wrap_qinpos = AHD_QIN_WRAP(qinpos); |
| wrap_qinfifonext = AHD_QIN_WRAP(ahd->qinfifonext); |
| if (wrap_qinfifonext >= wrap_qinpos) |
| return (wrap_qinfifonext - wrap_qinpos); |
| else |
| return (wrap_qinfifonext |
| + ARRAY_SIZE(ahd->qinfifo) - wrap_qinpos); |
| } |
| |
| static void |
| ahd_reset_cmds_pending(struct ahd_softc *ahd) |
| { |
| struct scb *scb; |
| ahd_mode_state saved_modes; |
| u_int pending_cmds; |
| |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_CCHAN, AHD_MODE_CCHAN); |
| |
| /* |
| * Don't count any commands as outstanding that the |
| * sequencer has already marked for completion. |
| */ |
| ahd_flush_qoutfifo(ahd); |
| |
| pending_cmds = 0; |
| LIST_FOREACH(scb, &ahd->pending_scbs, pending_links) { |
| pending_cmds++; |
| } |
| ahd_outw(ahd, CMDS_PENDING, pending_cmds - ahd_qinfifo_count(ahd)); |
| ahd_restore_modes(ahd, saved_modes); |
| ahd->flags &= ~AHD_UPDATE_PEND_CMDS; |
| } |
| |
| static void |
| ahd_done_with_status(struct ahd_softc *ahd, struct scb *scb, uint32_t status) |
| { |
| cam_status ostat; |
| cam_status cstat; |
| |
| ostat = ahd_get_transaction_status(scb); |
| if (ostat == CAM_REQ_INPROG) |
| ahd_set_transaction_status(scb, status); |
| cstat = ahd_get_transaction_status(scb); |
| if (cstat != CAM_REQ_CMP) |
| ahd_freeze_scb(scb); |
| ahd_done(ahd, scb); |
| } |
| |
| int |
| ahd_search_qinfifo(struct ahd_softc *ahd, int target, char channel, |
| int lun, u_int tag, role_t role, uint32_t status, |
| ahd_search_action action) |
| { |
| struct scb *scb; |
| struct scb *mk_msg_scb; |
| struct scb *prev_scb; |
| ahd_mode_state saved_modes; |
| u_int qinstart; |
| u_int qinpos; |
| u_int qintail; |
| u_int tid_next; |
| u_int tid_prev; |
| u_int scbid; |
| u_int seq_flags2; |
| u_int savedscbptr; |
| uint32_t busaddr; |
| int found; |
| int targets; |
| |
| /* Must be in CCHAN mode */ |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_CCHAN, AHD_MODE_CCHAN); |
| |
| /* |
| * Halt any pending SCB DMA. The sequencer will reinitiate |
| * this dma if the qinfifo is not empty once we unpause. |
| */ |
| if ((ahd_inb(ahd, CCSCBCTL) & (CCARREN|CCSCBEN|CCSCBDIR)) |
| == (CCARREN|CCSCBEN|CCSCBDIR)) { |
| ahd_outb(ahd, CCSCBCTL, |
| ahd_inb(ahd, CCSCBCTL) & ~(CCARREN|CCSCBEN)); |
| while ((ahd_inb(ahd, CCSCBCTL) & (CCARREN|CCSCBEN)) != 0) |
| ; |
| } |
| /* Determine sequencer's position in the qinfifo. */ |
| qintail = AHD_QIN_WRAP(ahd->qinfifonext); |
| qinstart = ahd_get_snscb_qoff(ahd); |
| qinpos = AHD_QIN_WRAP(qinstart); |
| found = 0; |
| prev_scb = NULL; |
| |
| if (action == SEARCH_PRINT) { |
| printk("qinstart = %d qinfifonext = %d\nQINFIFO:", |
| qinstart, ahd->qinfifonext); |
| } |
| |
| /* |
| * Start with an empty queue. Entries that are not chosen |
| * for removal will be re-added to the queue as we go. |
| */ |
| ahd->qinfifonext = qinstart; |
| busaddr = ahd_le32toh(ahd->next_queued_hscb->hscb_busaddr); |
| ahd_outl(ahd, NEXT_QUEUED_SCB_ADDR, busaddr); |
| |
| while (qinpos != qintail) { |
| scb = ahd_lookup_scb(ahd, ahd->qinfifo[qinpos]); |
| if (scb == NULL) { |
| printk("qinpos = %d, SCB index = %d\n", |
| qinpos, ahd->qinfifo[qinpos]); |
| panic("Loop 1\n"); |
| } |
| |
| if (ahd_match_scb(ahd, scb, target, channel, lun, tag, role)) { |
| /* |
| * We found an scb that needs to be acted on. |
| */ |
| found++; |
| switch (action) { |
| case SEARCH_COMPLETE: |
| if ((scb->flags & SCB_ACTIVE) == 0) |
| printk("Inactive SCB in qinfifo\n"); |
| ahd_done_with_status(ahd, scb, status); |
| /* FALLTHROUGH */ |
| case SEARCH_REMOVE: |
| break; |
| case SEARCH_PRINT: |
| printk(" 0x%x", ahd->qinfifo[qinpos]); |
| /* FALLTHROUGH */ |
| case SEARCH_COUNT: |
| ahd_qinfifo_requeue(ahd, prev_scb, scb); |
| prev_scb = scb; |
| break; |
| } |
| } else { |
| ahd_qinfifo_requeue(ahd, prev_scb, scb); |
| prev_scb = scb; |
| } |
| qinpos = AHD_QIN_WRAP(qinpos+1); |
| } |
| |
| ahd_set_hnscb_qoff(ahd, ahd->qinfifonext); |
| |
| if (action == SEARCH_PRINT) |
| printk("\nWAITING_TID_QUEUES:\n"); |
| |
| /* |
| * Search waiting for selection lists. We traverse the |
| * list of "their ids" waiting for selection and, if |
| * appropriate, traverse the SCBs of each "their id" |
| * looking for matches. |
| */ |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| seq_flags2 = ahd_inb(ahd, SEQ_FLAGS2); |
| if ((seq_flags2 & PENDING_MK_MESSAGE) != 0) { |
| scbid = ahd_inw(ahd, MK_MESSAGE_SCB); |
| mk_msg_scb = ahd_lookup_scb(ahd, scbid); |
| } else |
| mk_msg_scb = NULL; |
| savedscbptr = ahd_get_scbptr(ahd); |
| tid_next = ahd_inw(ahd, WAITING_TID_HEAD); |
| tid_prev = SCB_LIST_NULL; |
| targets = 0; |
| for (scbid = tid_next; !SCBID_IS_NULL(scbid); scbid = tid_next) { |
| u_int tid_head; |
| u_int tid_tail; |
| |
| targets++; |
| if (targets > AHD_NUM_TARGETS) |
| panic("TID LIST LOOP"); |
| |
| if (scbid >= ahd->scb_data.numscbs) { |
| printk("%s: Waiting TID List inconsistency. " |
| "SCB index == 0x%x, yet numscbs == 0x%x.", |
| ahd_name(ahd), scbid, ahd->scb_data.numscbs); |
| ahd_dump_card_state(ahd); |
| panic("for safety"); |
| } |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| printk("%s: SCB = 0x%x Not Active!\n", |
| ahd_name(ahd), scbid); |
| panic("Waiting TID List traversal\n"); |
| } |
| ahd_set_scbptr(ahd, scbid); |
| tid_next = ahd_inw_scbram(ahd, SCB_NEXT2); |
| if (ahd_match_scb(ahd, scb, target, channel, CAM_LUN_WILDCARD, |
| SCB_LIST_NULL, ROLE_UNKNOWN) == 0) { |
| tid_prev = scbid; |
| continue; |
| } |
| |
| /* |
| * We found a list of scbs that needs to be searched. |
| */ |
| if (action == SEARCH_PRINT) |
| printk(" %d ( ", SCB_GET_TARGET(ahd, scb)); |
| tid_head = scbid; |
| found += ahd_search_scb_list(ahd, target, channel, |
| lun, tag, role, status, |
| action, &tid_head, &tid_tail, |
| SCB_GET_TARGET(ahd, scb)); |
| /* |
| * Check any MK_MESSAGE SCB that is still waiting to |
| * enter this target's waiting for selection queue. |
| */ |
| if (mk_msg_scb != NULL |
| && ahd_match_scb(ahd, mk_msg_scb, target, channel, |
| lun, tag, role)) { |
| |
| /* |
| * We found an scb that needs to be acted on. |
| */ |
| found++; |
| switch (action) { |
| case SEARCH_COMPLETE: |
| if ((mk_msg_scb->flags & SCB_ACTIVE) == 0) |
| printk("Inactive SCB pending MK_MSG\n"); |
| ahd_done_with_status(ahd, mk_msg_scb, status); |
| /* FALLTHROUGH */ |
| case SEARCH_REMOVE: |
| { |
| u_int tail_offset; |
| |
| printk("Removing MK_MSG scb\n"); |
| |
| /* |
| * Reset our tail to the tail of the |
| * main per-target list. |
| */ |
| tail_offset = WAITING_SCB_TAILS |
| + (2 * SCB_GET_TARGET(ahd, mk_msg_scb)); |
| ahd_outw(ahd, tail_offset, tid_tail); |
| |
| seq_flags2 &= ~PENDING_MK_MESSAGE; |
| ahd_outb(ahd, SEQ_FLAGS2, seq_flags2); |
| ahd_outw(ahd, CMDS_PENDING, |
| ahd_inw(ahd, CMDS_PENDING)-1); |
| mk_msg_scb = NULL; |
| break; |
| } |
| case SEARCH_PRINT: |
| printk(" 0x%x", SCB_GET_TAG(scb)); |
| /* FALLTHROUGH */ |
| case SEARCH_COUNT: |
| break; |
| } |
| } |
| |
| if (mk_msg_scb != NULL |
| && SCBID_IS_NULL(tid_head) |
| && ahd_match_scb(ahd, scb, target, channel, CAM_LUN_WILDCARD, |
| SCB_LIST_NULL, ROLE_UNKNOWN)) { |
| |
| /* |
| * When removing the last SCB for a target |
| * queue with a pending MK_MESSAGE scb, we |
| * must queue the MK_MESSAGE scb. |
| */ |
| printk("Queueing mk_msg_scb\n"); |
| tid_head = ahd_inw(ahd, MK_MESSAGE_SCB); |
| seq_flags2 &= ~PENDING_MK_MESSAGE; |
| ahd_outb(ahd, SEQ_FLAGS2, seq_flags2); |
| mk_msg_scb = NULL; |
| } |
| if (tid_head != scbid) |
| ahd_stitch_tid_list(ahd, tid_prev, tid_head, tid_next); |
| if (!SCBID_IS_NULL(tid_head)) |
| tid_prev = tid_head; |
| if (action == SEARCH_PRINT) |
| printk(")\n"); |
| } |
| |
| /* Restore saved state. */ |
| ahd_set_scbptr(ahd, savedscbptr); |
| ahd_restore_modes(ahd, saved_modes); |
| return (found); |
| } |
| |
| static int |
| ahd_search_scb_list(struct ahd_softc *ahd, int target, char channel, |
| int lun, u_int tag, role_t role, uint32_t status, |
| ahd_search_action action, u_int *list_head, |
| u_int *list_tail, u_int tid) |
| { |
| struct scb *scb; |
| u_int scbid; |
| u_int next; |
| u_int prev; |
| int found; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| found = 0; |
| prev = SCB_LIST_NULL; |
| next = *list_head; |
| *list_tail = SCB_LIST_NULL; |
| for (scbid = next; !SCBID_IS_NULL(scbid); scbid = next) { |
| if (scbid >= ahd->scb_data.numscbs) { |
| printk("%s:SCB List inconsistency. " |
| "SCB == 0x%x, yet numscbs == 0x%x.", |
| ahd_name(ahd), scbid, ahd->scb_data.numscbs); |
| ahd_dump_card_state(ahd); |
| panic("for safety"); |
| } |
| scb = ahd_lookup_scb(ahd, scbid); |
| if (scb == NULL) { |
| printk("%s: SCB = %d Not Active!\n", |
| ahd_name(ahd), scbid); |
| panic("Waiting List traversal\n"); |
| } |
| ahd_set_scbptr(ahd, scbid); |
| *list_tail = scbid; |
| next = ahd_inw_scbram(ahd, SCB_NEXT); |
| if (ahd_match_scb(ahd, scb, target, channel, |
| lun, SCB_LIST_NULL, role) == 0) { |
| prev = scbid; |
| continue; |
| } |
| found++; |
| switch (action) { |
| case SEARCH_COMPLETE: |
| if ((scb->flags & SCB_ACTIVE) == 0) |
| printk("Inactive SCB in Waiting List\n"); |
| ahd_done_with_status(ahd, scb, status); |
| /* FALLTHROUGH */ |
| case SEARCH_REMOVE: |
| ahd_rem_wscb(ahd, scbid, prev, next, tid); |
| *list_tail = prev; |
| if (SCBID_IS_NULL(prev)) |
| *list_head = next; |
| break; |
| case SEARCH_PRINT: |
| printk("0x%x ", scbid); |
| case SEARCH_COUNT: |
| prev = scbid; |
| break; |
| } |
| if (found > AHD_SCB_MAX) |
| panic("SCB LIST LOOP"); |
| } |
| if (action == SEARCH_COMPLETE |
| || action == SEARCH_REMOVE) |
| ahd_outw(ahd, CMDS_PENDING, ahd_inw(ahd, CMDS_PENDING) - found); |
| return (found); |
| } |
| |
| static void |
| ahd_stitch_tid_list(struct ahd_softc *ahd, u_int tid_prev, |
| u_int tid_cur, u_int tid_next) |
| { |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| |
| if (SCBID_IS_NULL(tid_cur)) { |
| |
| /* Bypass current TID list */ |
| if (SCBID_IS_NULL(tid_prev)) { |
| ahd_outw(ahd, WAITING_TID_HEAD, tid_next); |
| } else { |
| ahd_set_scbptr(ahd, tid_prev); |
| ahd_outw(ahd, SCB_NEXT2, tid_next); |
| } |
| if (SCBID_IS_NULL(tid_next)) |
| ahd_outw(ahd, WAITING_TID_TAIL, tid_prev); |
| } else { |
| |
| /* Stitch through tid_cur */ |
| if (SCBID_IS_NULL(tid_prev)) { |
| ahd_outw(ahd, WAITING_TID_HEAD, tid_cur); |
| } else { |
| ahd_set_scbptr(ahd, tid_prev); |
| ahd_outw(ahd, SCB_NEXT2, tid_cur); |
| } |
| ahd_set_scbptr(ahd, tid_cur); |
| ahd_outw(ahd, SCB_NEXT2, tid_next); |
| |
| if (SCBID_IS_NULL(tid_next)) |
| ahd_outw(ahd, WAITING_TID_TAIL, tid_cur); |
| } |
| } |
| |
| /* |
| * Manipulate the waiting for selection list and return the |
| * scb that follows the one that we remove. |
| */ |
| static u_int |
| ahd_rem_wscb(struct ahd_softc *ahd, u_int scbid, |
| u_int prev, u_int next, u_int tid) |
| { |
| u_int tail_offset; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| if (!SCBID_IS_NULL(prev)) { |
| ahd_set_scbptr(ahd, prev); |
| ahd_outw(ahd, SCB_NEXT, next); |
| } |
| |
| /* |
| * SCBs that have MK_MESSAGE set in them may |
| * cause the tail pointer to be updated without |
| * setting the next pointer of the previous tail. |
| * Only clear the tail if the removed SCB was |
| * the tail. |
| */ |
| tail_offset = WAITING_SCB_TAILS + (2 * tid); |
| if (SCBID_IS_NULL(next) |
| && ahd_inw(ahd, tail_offset) == scbid) |
| ahd_outw(ahd, tail_offset, prev); |
| |
| ahd_add_scb_to_free_list(ahd, scbid); |
| return (next); |
| } |
| |
| /* |
| * Add the SCB as selected by SCBPTR onto the on chip list of |
| * free hardware SCBs. This list is empty/unused if we are not |
| * performing SCB paging. |
| */ |
| static void |
| ahd_add_scb_to_free_list(struct ahd_softc *ahd, u_int scbid) |
| { |
| /* XXX Need some other mechanism to designate "free". */ |
| /* |
| * Invalidate the tag so that our abort |
| * routines don't think it's active. |
| ahd_outb(ahd, SCB_TAG, SCB_LIST_NULL); |
| */ |
| } |
| |
| /******************************** Error Handling ******************************/ |
| /* |
| * Abort all SCBs that match the given description (target/channel/lun/tag), |
| * setting their status to the passed in status if the status has not already |
| * been modified from CAM_REQ_INPROG. This routine assumes that the sequencer |
| * is paused before it is called. |
| */ |
| static int |
| ahd_abort_scbs(struct ahd_softc *ahd, int target, char channel, |
| int lun, u_int tag, role_t role, uint32_t status) |
| { |
| struct scb *scbp; |
| struct scb *scbp_next; |
| u_int i, j; |
| u_int maxtarget; |
| u_int minlun; |
| u_int maxlun; |
| int found; |
| ahd_mode_state saved_modes; |
| |
| /* restore this when we're done */ |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| |
| found = ahd_search_qinfifo(ahd, target, channel, lun, SCB_LIST_NULL, |
| role, CAM_REQUEUE_REQ, SEARCH_COMPLETE); |
| |
| /* |
| * Clean out the busy target table for any untagged commands. |
| */ |
| i = 0; |
| maxtarget = 16; |
| if (target != CAM_TARGET_WILDCARD) { |
| i = target; |
| if (channel == 'B') |
| i += 8; |
| maxtarget = i + 1; |
| } |
| |
| if (lun == CAM_LUN_WILDCARD) { |
| minlun = 0; |
| maxlun = AHD_NUM_LUNS_NONPKT; |
| } else if (lun >= AHD_NUM_LUNS_NONPKT) { |
| minlun = maxlun = 0; |
| } else { |
| minlun = lun; |
| maxlun = lun + 1; |
| } |
| |
| if (role != ROLE_TARGET) { |
| for (;i < maxtarget; i++) { |
| for (j = minlun;j < maxlun; j++) { |
| u_int scbid; |
| u_int tcl; |
| |
| tcl = BUILD_TCL_RAW(i, 'A', j); |
| scbid = ahd_find_busy_tcl(ahd, tcl); |
| scbp = ahd_lookup_scb(ahd, scbid); |
| if (scbp == NULL |
| || ahd_match_scb(ahd, scbp, target, channel, |
| lun, tag, role) == 0) |
| continue; |
| ahd_unbusy_tcl(ahd, BUILD_TCL_RAW(i, 'A', j)); |
| } |
| } |
| } |
| |
| /* |
| * Don't abort commands that have already completed, |
| * but haven't quite made it up to the host yet. |
| */ |
| ahd_flush_qoutfifo(ahd); |
| |
| /* |
| * Go through the pending CCB list and look for |
| * commands for this target that are still active. |
| * These are other tagged commands that were |
| * disconnected when the reset occurred. |
| */ |
| scbp_next = LIST_FIRST(&ahd->pending_scbs); |
| while (scbp_next != NULL) { |
| scbp = scbp_next; |
| scbp_next = LIST_NEXT(scbp, pending_links); |
| if (ahd_match_scb(ahd, scbp, target, channel, lun, tag, role)) { |
| cam_status ostat; |
| |
| ostat = ahd_get_transaction_status(scbp); |
| if (ostat == CAM_REQ_INPROG) |
| ahd_set_transaction_status(scbp, status); |
| if (ahd_get_transaction_status(scbp) != CAM_REQ_CMP) |
| ahd_freeze_scb(scbp); |
| if ((scbp->flags & SCB_ACTIVE) == 0) |
| printk("Inactive SCB on pending list\n"); |
| ahd_done(ahd, scbp); |
| found++; |
| } |
| } |
| ahd_restore_modes(ahd, saved_modes); |
| ahd_platform_abort_scbs(ahd, target, channel, lun, tag, role, status); |
| ahd->flags |= AHD_UPDATE_PEND_CMDS; |
| return found; |
| } |
| |
| static void |
| ahd_reset_current_bus(struct ahd_softc *ahd) |
| { |
| uint8_t scsiseq; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| ahd_outb(ahd, SIMODE1, ahd_inb(ahd, SIMODE1) & ~ENSCSIRST); |
| scsiseq = ahd_inb(ahd, SCSISEQ0) & ~(ENSELO|ENARBO|SCSIRSTO); |
| ahd_outb(ahd, SCSISEQ0, scsiseq | SCSIRSTO); |
| ahd_flush_device_writes(ahd); |
| ahd_delay(AHD_BUSRESET_DELAY); |
| /* Turn off the bus reset */ |
| ahd_outb(ahd, SCSISEQ0, scsiseq); |
| ahd_flush_device_writes(ahd); |
| ahd_delay(AHD_BUSRESET_DELAY); |
| if ((ahd->bugs & AHD_SCSIRST_BUG) != 0) { |
| /* |
| * 2A Razor #474 |
| * Certain chip state is not cleared for |
| * SCSI bus resets that we initiate, so |
| * we must reset the chip. |
| */ |
| ahd_reset(ahd, /*reinit*/TRUE); |
| ahd_intr_enable(ahd, /*enable*/TRUE); |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| } |
| |
| ahd_clear_intstat(ahd); |
| } |
| |
| int |
| ahd_reset_channel(struct ahd_softc *ahd, char channel, int initiate_reset) |
| { |
| struct ahd_devinfo caminfo; |
| u_int initiator; |
| u_int target; |
| u_int max_scsiid; |
| int found; |
| u_int fifo; |
| u_int next_fifo; |
| uint8_t scsiseq; |
| |
| /* |
| * Check if the last bus reset is cleared |
| */ |
| if (ahd->flags & AHD_BUS_RESET_ACTIVE) { |
| printk("%s: bus reset still active\n", |
| ahd_name(ahd)); |
| return 0; |
| } |
| ahd->flags |= AHD_BUS_RESET_ACTIVE; |
| |
| ahd->pending_device = NULL; |
| |
| ahd_compile_devinfo(&caminfo, |
| CAM_TARGET_WILDCARD, |
| CAM_TARGET_WILDCARD, |
| CAM_LUN_WILDCARD, |
| channel, ROLE_UNKNOWN); |
| ahd_pause(ahd); |
| |
| /* Make sure the sequencer is in a safe location. */ |
| ahd_clear_critical_section(ahd); |
| |
| /* |
| * Run our command complete fifos to ensure that we perform |
| * completion processing on any commands that 'completed' |
| * before the reset occurred. |
| */ |
| ahd_run_qoutfifo(ahd); |
| #ifdef AHD_TARGET_MODE |
| if ((ahd->flags & AHD_TARGETROLE) != 0) { |
| ahd_run_tqinfifo(ahd, /*paused*/TRUE); |
| } |
| #endif |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| |
| /* |
| * Disable selections so no automatic hardware |
| * functions will modify chip state. |
| */ |
| ahd_outb(ahd, SCSISEQ0, 0); |
| ahd_outb(ahd, SCSISEQ1, 0); |
| |
| /* |
| * Safely shut down our DMA engines. Always start with |
| * the FIFO that is not currently active (if any are |
| * actively connected). |
| */ |
| next_fifo = fifo = ahd_inb(ahd, DFFSTAT) & CURRFIFO; |
| if (next_fifo > CURRFIFO_1) |
| /* If disconneced, arbitrarily start with FIFO1. */ |
| next_fifo = fifo = 0; |
| do { |
| next_fifo ^= CURRFIFO_1; |
| ahd_set_modes(ahd, next_fifo, next_fifo); |
| ahd_outb(ahd, DFCNTRL, |
| ahd_inb(ahd, DFCNTRL) & ~(SCSIEN|HDMAEN)); |
| while ((ahd_inb(ahd, DFCNTRL) & HDMAENACK) != 0) |
| ahd_delay(10); |
| /* |
| * Set CURRFIFO to the now inactive channel. |
| */ |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| ahd_outb(ahd, DFFSTAT, next_fifo); |
| } while (next_fifo != fifo); |
| |
| /* |
| * Reset the bus if we are initiating this reset |
| */ |
| ahd_clear_msg_state(ahd); |
| ahd_outb(ahd, SIMODE1, |
| ahd_inb(ahd, SIMODE1) & ~(ENBUSFREE|ENSCSIRST)); |
| |
| if (initiate_reset) |
| ahd_reset_current_bus(ahd); |
| |
| ahd_clear_intstat(ahd); |
| |
| /* |
| * Clean up all the state information for the |
| * pending transactions on this bus. |
| */ |
| found = ahd_abort_scbs(ahd, CAM_TARGET_WILDCARD, channel, |
| CAM_LUN_WILDCARD, SCB_LIST_NULL, |
| ROLE_UNKNOWN, CAM_SCSI_BUS_RESET); |
| |
| /* |
| * Cleanup anything left in the FIFOs. |
| */ |
| ahd_clear_fifo(ahd, 0); |
| ahd_clear_fifo(ahd, 1); |
| |
| /* |
| * Clear SCSI interrupt status |
| */ |
| ahd_outb(ahd, CLRSINT1, CLRSCSIRSTI); |
| |
| /* |
| * Reenable selections |
| */ |
| ahd_outb(ahd, SIMODE1, ahd_inb(ahd, SIMODE1) | ENSCSIRST); |
| scsiseq = ahd_inb(ahd, SCSISEQ_TEMPLATE); |
| ahd_outb(ahd, SCSISEQ1, scsiseq & (ENSELI|ENRSELI|ENAUTOATNP)); |
| |
| max_scsiid = (ahd->features & AHD_WIDE) ? 15 : 7; |
| #ifdef AHD_TARGET_MODE |
| /* |
| * Send an immediate notify ccb to all target more peripheral |
| * drivers affected by this action. |
| */ |
| for (target = 0; target <= max_scsiid; target++) { |
| struct ahd_tmode_tstate* tstate; |
| u_int lun; |
| |
| tstate = ahd->enabled_targets[target]; |
| if (tstate == NULL) |
| continue; |
| for (lun = 0; lun < AHD_NUM_LUNS; lun++) { |
| struct ahd_tmode_lstate* lstate; |
| |
| lstate = tstate->enabled_luns[lun]; |
| if (lstate == NULL) |
| continue; |
| |
| ahd_queue_lstate_event(ahd, lstate, CAM_TARGET_WILDCARD, |
| EVENT_TYPE_BUS_RESET, /*arg*/0); |
| ahd_send_lstate_events(ahd, lstate); |
| } |
| } |
| #endif |
| /* |
| * Revert to async/narrow transfers until we renegotiate. |
| */ |
| for (target = 0; target <= max_scsiid; target++) { |
| |
| if (ahd->enabled_targets[target] == NULL) |
| continue; |
| for (initiator = 0; initiator <= max_scsiid; initiator++) { |
| struct ahd_devinfo devinfo; |
| |
| ahd_compile_devinfo(&devinfo, target, initiator, |
| CAM_LUN_WILDCARD, |
| 'A', ROLE_UNKNOWN); |
| ahd_set_width(ahd, &devinfo, MSG_EXT_WDTR_BUS_8_BIT, |
| AHD_TRANS_CUR, /*paused*/TRUE); |
| ahd_set_syncrate(ahd, &devinfo, /*period*/0, |
| /*offset*/0, /*ppr_options*/0, |
| AHD_TRANS_CUR, /*paused*/TRUE); |
| } |
| } |
| |
| /* Notify the XPT that a bus reset occurred */ |
| ahd_send_async(ahd, caminfo.channel, CAM_TARGET_WILDCARD, |
| CAM_LUN_WILDCARD, AC_BUS_RESET); |
| |
| ahd_restart(ahd); |
| |
| return (found); |
| } |
| |
| /**************************** Statistics Processing ***************************/ |
| static void |
| ahd_stat_timer(void *arg) |
| { |
| struct ahd_softc *ahd = arg; |
| u_long s; |
| int enint_coal; |
| |
| ahd_lock(ahd, &s); |
| |
| enint_coal = ahd->hs_mailbox & ENINT_COALESCE; |
| if (ahd->cmdcmplt_total > ahd->int_coalescing_threshold) |
| enint_coal |= ENINT_COALESCE; |
| else if (ahd->cmdcmplt_total < ahd->int_coalescing_stop_threshold) |
| enint_coal &= ~ENINT_COALESCE; |
| |
| if (enint_coal != (ahd->hs_mailbox & ENINT_COALESCE)) { |
| ahd_enable_coalescing(ahd, enint_coal); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_INT_COALESCING) != 0) |
| printk("%s: Interrupt coalescing " |
| "now %sabled. Cmds %d\n", |
| ahd_name(ahd), |
| (enint_coal & ENINT_COALESCE) ? "en" : "dis", |
| ahd->cmdcmplt_total); |
| #endif |
| } |
| |
| ahd->cmdcmplt_bucket = (ahd->cmdcmplt_bucket+1) & (AHD_STAT_BUCKETS-1); |
| ahd->cmdcmplt_total -= ahd->cmdcmplt_counts[ahd->cmdcmplt_bucket]; |
| ahd->cmdcmplt_counts[ahd->cmdcmplt_bucket] = 0; |
| ahd_timer_reset(&ahd->stat_timer, AHD_STAT_UPDATE_US, |
| ahd_stat_timer, ahd); |
| ahd_unlock(ahd, &s); |
| } |
| |
| /****************************** Status Processing *****************************/ |
| |
| static void |
| ahd_handle_scsi_status(struct ahd_softc *ahd, struct scb *scb) |
| { |
| struct hardware_scb *hscb; |
| int paused; |
| |
| /* |
| * The sequencer freezes its select-out queue |
| * anytime a SCSI status error occurs. We must |
| * handle the error and increment our qfreeze count |
| * to allow the sequencer to continue. We don't |
| * bother clearing critical sections here since all |
| * operations are on data structures that the sequencer |
| * is not touching once the queue is frozen. |
| */ |
| hscb = scb->hscb; |
| |
| if (ahd_is_paused(ahd)) { |
| paused = 1; |
| } else { |
| paused = 0; |
| ahd_pause(ahd); |
| } |
| |
| /* Freeze the queue until the client sees the error. */ |
| ahd_freeze_devq(ahd, scb); |
| ahd_freeze_scb(scb); |
| ahd->qfreeze_cnt++; |
| ahd_outw(ahd, KERNEL_QFREEZE_COUNT, ahd->qfreeze_cnt); |
| |
| if (paused == 0) |
| ahd_unpause(ahd); |
| |
| /* Don't want to clobber the original sense code */ |
| if ((scb->flags & SCB_SENSE) != 0) { |
| /* |
| * Clear the SCB_SENSE Flag and perform |
| * a normal command completion. |
| */ |
| scb->flags &= ~SCB_SENSE; |
| ahd_set_transaction_status(scb, CAM_AUTOSENSE_FAIL); |
| ahd_done(ahd, scb); |
| return; |
| } |
| ahd_set_transaction_status(scb, CAM_SCSI_STATUS_ERROR); |
| ahd_set_scsi_status(scb, hscb->shared_data.istatus.scsi_status); |
| switch (hscb->shared_data.istatus.scsi_status) { |
| case STATUS_PKT_SENSE: |
| { |
| struct scsi_status_iu_header *siu; |
| |
| ahd_sync_sense(ahd, scb, BUS_DMASYNC_POSTREAD); |
| siu = (struct scsi_status_iu_header *)scb->sense_data; |
| ahd_set_scsi_status(scb, siu->status); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_SENSE) != 0) { |
| ahd_print_path(ahd, scb); |
| printk("SCB 0x%x Received PKT Status of 0x%x\n", |
| SCB_GET_TAG(scb), siu->status); |
| printk("\tflags = 0x%x, sense len = 0x%x, " |
| "pktfail = 0x%x\n", |
| siu->flags, scsi_4btoul(siu->sense_length), |
| scsi_4btoul(siu->pkt_failures_length)); |
| } |
| #endif |
| if ((siu->flags & SIU_RSPVALID) != 0) { |
| ahd_print_path(ahd, scb); |
| if (scsi_4btoul(siu->pkt_failures_length) < 4) { |
| printk("Unable to parse pkt_failures\n"); |
| } else { |
| |
| switch (SIU_PKTFAIL_CODE(siu)) { |
| case SIU_PFC_NONE: |
| printk("No packet failure found\n"); |
| break; |
| case SIU_PFC_CIU_FIELDS_INVALID: |
| printk("Invalid Command IU Field\n"); |
| break; |
| case SIU_PFC_TMF_NOT_SUPPORTED: |
| printk("TMF not supported\n"); |
| break; |
| case SIU_PFC_TMF_FAILED: |
| printk("TMF failed\n"); |
| break; |
| case SIU_PFC_INVALID_TYPE_CODE: |
| printk("Invalid L_Q Type code\n"); |
| break; |
| case SIU_PFC_ILLEGAL_REQUEST: |
| printk("Illegal request\n"); |
| default: |
| break; |
| } |
| } |
| if (siu->status == SCSI_STATUS_OK) |
| ahd_set_transaction_status(scb, |
| CAM_REQ_CMP_ERR); |
| } |
| if ((siu->flags & SIU_SNSVALID) != 0) { |
| scb->flags |= SCB_PKT_SENSE; |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_SENSE) != 0) |
| printk("Sense data available\n"); |
| #endif |
| } |
| ahd_done(ahd, scb); |
| break; |
| } |
| case SCSI_STATUS_CMD_TERMINATED: |
| case SCSI_STATUS_CHECK_COND: |
| { |
| struct ahd_devinfo devinfo; |
| struct ahd_dma_seg *sg; |
| struct scsi_sense *sc; |
| struct ahd_initiator_tinfo *targ_info; |
| struct ahd_tmode_tstate *tstate; |
| struct ahd_transinfo *tinfo; |
| #ifdef AHD_DEBUG |
| if (ahd_debug & AHD_SHOW_SENSE) { |
| ahd_print_path(ahd, scb); |
| printk("SCB %d: requests Check Status\n", |
| SCB_GET_TAG(scb)); |
| } |
| #endif |
| |
| if (ahd_perform_autosense(scb) == 0) |
| break; |
| |
| ahd_compile_devinfo(&devinfo, SCB_GET_OUR_ID(scb), |
| SCB_GET_TARGET(ahd, scb), |
| SCB_GET_LUN(scb), |
| SCB_GET_CHANNEL(ahd, scb), |
| ROLE_INITIATOR); |
| targ_info = ahd_fetch_transinfo(ahd, |
| devinfo.channel, |
| devinfo.our_scsiid, |
| devinfo.target, |
| &tstate); |
| tinfo = &targ_info->curr; |
| sg = scb->sg_list; |
| sc = (struct scsi_sense *)hscb->shared_data.idata.cdb; |
| /* |
| * Save off the residual if there is one. |
| */ |
| ahd_update_residual(ahd, scb); |
| #ifdef AHD_DEBUG |
| if (ahd_debug & AHD_SHOW_SENSE) { |
| ahd_print_path(ahd, scb); |
| printk("Sending Sense\n"); |
| } |
| #endif |
| scb->sg_count = 0; |
| sg = ahd_sg_setup(ahd, scb, sg, ahd_get_sense_bufaddr(ahd, scb), |
| ahd_get_sense_bufsize(ahd, scb), |
| /*last*/TRUE); |
| sc->opcode = REQUEST_SENSE; |
| sc->byte2 = 0; |
| if (tinfo->protocol_version <= SCSI_REV_2 |
| && SCB_GET_LUN(scb) < 8) |
| sc->byte2 = SCB_GET_LUN(scb) << 5; |
| sc->unused[0] = 0; |
| sc->unused[1] = 0; |
| sc->length = ahd_get_sense_bufsize(ahd, scb); |
| sc->control = 0; |
| |
| /* |
| * We can't allow the target to disconnect. |
| * This will be an untagged transaction and |
| * having the target disconnect will make this |
| * transaction indestinguishable from outstanding |
| * tagged transactions. |
| */ |
| hscb->control = 0; |
| |
| /* |
| * This request sense could be because the |
| * the device lost power or in some other |
| * way has lost our transfer negotiations. |
| * Renegotiate if appropriate. Unit attention |
| * errors will be reported before any data |
| * phases occur. |
| */ |
| if (ahd_get_residual(scb) == ahd_get_transfer_length(scb)) { |
| ahd_update_neg_request(ahd, &devinfo, |
| tstate, targ_info, |
| AHD_NEG_IF_NON_ASYNC); |
| } |
| if (tstate->auto_negotiate & devinfo.target_mask) { |
| hscb->control |= MK_MESSAGE; |
| scb->flags &= |
| ~(SCB_NEGOTIATE|SCB_ABORT|SCB_DEVICE_RESET); |
| scb->flags |= SCB_AUTO_NEGOTIATE; |
| } |
| hscb->cdb_len = sizeof(*sc); |
| ahd_setup_data_scb(ahd, scb); |
| scb->flags |= SCB_SENSE; |
| ahd_queue_scb(ahd, scb); |
| break; |
| } |
| case SCSI_STATUS_OK: |
| printk("%s: Interrupted for status of 0???\n", |
| ahd_name(ahd)); |
| /* FALLTHROUGH */ |
| default: |
| ahd_done(ahd, scb); |
| break; |
| } |
| } |
| |
| static void |
| ahd_handle_scb_status(struct ahd_softc *ahd, struct scb *scb) |
| { |
| if (scb->hscb->shared_data.istatus.scsi_status != 0) { |
| ahd_handle_scsi_status(ahd, scb); |
| } else { |
| ahd_calc_residual(ahd, scb); |
| ahd_done(ahd, scb); |
| } |
| } |
| |
| /* |
| * Calculate the residual for a just completed SCB. |
| */ |
| static void |
| ahd_calc_residual(struct ahd_softc *ahd, struct scb *scb) |
| { |
| struct hardware_scb *hscb; |
| struct initiator_status *spkt; |
| uint32_t sgptr; |
| uint32_t resid_sgptr; |
| uint32_t resid; |
| |
| /* |
| * 5 cases. |
| * 1) No residual. |
| * SG_STATUS_VALID clear in sgptr. |
| * 2) Transferless command |
| * 3) Never performed any transfers. |
| * sgptr has SG_FULL_RESID set. |
| * 4) No residual but target did not |
| * save data pointers after the |
| * last transfer, so sgptr was |
| * never updated. |
| * 5) We have a partial residual. |
| * Use residual_sgptr to determine |
| * where we are. |
| */ |
| |
| hscb = scb->hscb; |
| sgptr = ahd_le32toh(hscb->sgptr); |
| if ((sgptr & SG_STATUS_VALID) == 0) |
| /* Case 1 */ |
| return; |
| sgptr &= ~SG_STATUS_VALID; |
| |
| if ((sgptr & SG_LIST_NULL) != 0) |
| /* Case 2 */ |
| return; |
| |
| /* |
| * Residual fields are the same in both |
| * target and initiator status packets, |
| * so we can always use the initiator fields |
| * regardless of the role for this SCB. |
| */ |
| spkt = &hscb->shared_data.istatus; |
| resid_sgptr = ahd_le32toh(spkt->residual_sgptr); |
| if ((sgptr & SG_FULL_RESID) != 0) { |
| /* Case 3 */ |
| resid = ahd_get_transfer_length(scb); |
| } else if ((resid_sgptr & SG_LIST_NULL) != 0) { |
| /* Case 4 */ |
| return; |
| } else if ((resid_sgptr & SG_OVERRUN_RESID) != 0) { |
| ahd_print_path(ahd, scb); |
| printk("data overrun detected Tag == 0x%x.\n", |
| SCB_GET_TAG(scb)); |
| ahd_freeze_devq(ahd, scb); |
| ahd_set_transaction_status(scb, CAM_DATA_RUN_ERR); |
| ahd_freeze_scb(scb); |
| return; |
| } else if ((resid_sgptr & ~SG_PTR_MASK) != 0) { |
| panic("Bogus resid sgptr value 0x%x\n", resid_sgptr); |
| /* NOTREACHED */ |
| } else { |
| struct ahd_dma_seg *sg; |
| |
| /* |
| * Remainder of the SG where the transfer |
| * stopped. |
| */ |
| resid = ahd_le32toh(spkt->residual_datacnt) & AHD_SG_LEN_MASK; |
| sg = ahd_sg_bus_to_virt(ahd, scb, resid_sgptr & SG_PTR_MASK); |
| |
| /* The residual sg_ptr always points to the next sg */ |
| sg--; |
| |
| /* |
| * Add up the contents of all residual |
| * SG segments that are after the SG where |
| * the transfer stopped. |
| */ |
| while ((ahd_le32toh(sg->len) & AHD_DMA_LAST_SEG) == 0) { |
| sg++; |
| resid += ahd_le32toh(sg->len) & AHD_SG_LEN_MASK; |
| } |
| } |
| if ((scb->flags & SCB_SENSE) == 0) |
| ahd_set_residual(scb, resid); |
| else |
| ahd_set_sense_residual(scb, resid); |
| |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_MISC) != 0) { |
| ahd_print_path(ahd, scb); |
| printk("Handled %sResidual of %d bytes\n", |
| (scb->flags & SCB_SENSE) ? "Sense " : "", resid); |
| } |
| #endif |
| } |
| |
| /******************************* Target Mode **********************************/ |
| #ifdef AHD_TARGET_MODE |
| /* |
| * Add a target mode event to this lun's queue |
| */ |
| static void |
| ahd_queue_lstate_event(struct ahd_softc *ahd, struct ahd_tmode_lstate *lstate, |
| u_int initiator_id, u_int event_type, u_int event_arg) |
| { |
| struct ahd_tmode_event *event; |
| int pending; |
| |
| xpt_freeze_devq(lstate->path, /*count*/1); |
| if (lstate->event_w_idx >= lstate->event_r_idx) |
| pending = lstate->event_w_idx - lstate->event_r_idx; |
| else |
| pending = AHD_TMODE_EVENT_BUFFER_SIZE + 1 |
| - (lstate->event_r_idx - lstate->event_w_idx); |
| |
| if (event_type == EVENT_TYPE_BUS_RESET |
| || event_type == MSG_BUS_DEV_RESET) { |
| /* |
| * Any earlier events are irrelevant, so reset our buffer. |
| * This has the effect of allowing us to deal with reset |
| * floods (an external device holding down the reset line) |
| * without losing the event that is really interesting. |
| */ |
| lstate->event_r_idx = 0; |
| lstate->event_w_idx = 0; |
| xpt_release_devq(lstate->path, pending, /*runqueue*/FALSE); |
| } |
| |
| if (pending == AHD_TMODE_EVENT_BUFFER_SIZE) { |
| xpt_print_path(lstate->path); |
| printk("immediate event %x:%x lost\n", |
| lstate->event_buffer[lstate->event_r_idx].event_type, |
| lstate->event_buffer[lstate->event_r_idx].event_arg); |
| lstate->event_r_idx++; |
| if (lstate->event_r_idx == AHD_TMODE_EVENT_BUFFER_SIZE) |
| lstate->event_r_idx = 0; |
| xpt_release_devq(lstate->path, /*count*/1, /*runqueue*/FALSE); |
| } |
| |
| event = &lstate->event_buffer[lstate->event_w_idx]; |
| event->initiator_id = initiator_id; |
| event->event_type = event_type; |
| event->event_arg = event_arg; |
| lstate->event_w_idx++; |
| if (lstate->event_w_idx == AHD_TMODE_EVENT_BUFFER_SIZE) |
| lstate->event_w_idx = 0; |
| } |
| |
| /* |
| * Send any target mode events queued up waiting |
| * for immediate notify resources. |
| */ |
| void |
| ahd_send_lstate_events(struct ahd_softc *ahd, struct ahd_tmode_lstate *lstate) |
| { |
| struct ccb_hdr *ccbh; |
| struct ccb_immed_notify *inot; |
| |
| while (lstate->event_r_idx != lstate->event_w_idx |
| && (ccbh = SLIST_FIRST(&lstate->immed_notifies)) != NULL) { |
| struct ahd_tmode_event *event; |
| |
| event = &lstate->event_buffer[lstate->event_r_idx]; |
| SLIST_REMOVE_HEAD(&lstate->immed_notifies, sim_links.sle); |
| inot = (struct ccb_immed_notify *)ccbh; |
| switch (event->event_type) { |
| case EVENT_TYPE_BUS_RESET: |
| ccbh->status = CAM_SCSI_BUS_RESET|CAM_DEV_QFRZN; |
| break; |
| default: |
| ccbh->status = CAM_MESSAGE_RECV|CAM_DEV_QFRZN; |
| inot->message_args[0] = event->event_type; |
| inot->message_args[1] = event->event_arg; |
| break; |
| } |
| inot->initiator_id = event->initiator_id; |
| inot->sense_len = 0; |
| xpt_done((union ccb *)inot); |
| lstate->event_r_idx++; |
| if (lstate->event_r_idx == AHD_TMODE_EVENT_BUFFER_SIZE) |
| lstate->event_r_idx = 0; |
| } |
| } |
| #endif |
| |
| /******************** Sequencer Program Patching/Download *********************/ |
| |
| #ifdef AHD_DUMP_SEQ |
| void |
| ahd_dumpseq(struct ahd_softc* ahd) |
| { |
| int i; |
| int max_prog; |
| |
| max_prog = 2048; |
| |
| ahd_outb(ahd, SEQCTL0, PERRORDIS|FAILDIS|FASTMODE|LOADRAM); |
| ahd_outw(ahd, PRGMCNT, 0); |
| for (i = 0; i < max_prog; i++) { |
| uint8_t ins_bytes[4]; |
| |
| ahd_insb(ahd, SEQRAM, ins_bytes, 4); |
| printk("0x%08x\n", ins_bytes[0] << 24 |
| | ins_bytes[1] << 16 |
| | ins_bytes[2] << 8 |
| | ins_bytes[3]); |
| } |
| } |
| #endif |
| |
| static void |
| ahd_loadseq(struct ahd_softc *ahd) |
| { |
| struct cs cs_table[num_critical_sections]; |
| u_int begin_set[num_critical_sections]; |
| u_int end_set[num_critical_sections]; |
| const struct patch *cur_patch; |
| u_int cs_count; |
| u_int cur_cs; |
| u_int i; |
| int downloaded; |
| u_int skip_addr; |
| u_int sg_prefetch_cnt; |
| u_int sg_prefetch_cnt_limit; |
| u_int sg_prefetch_align; |
| u_int sg_size; |
| u_int cacheline_mask; |
| uint8_t download_consts[DOWNLOAD_CONST_COUNT]; |
| |
| if (bootverbose) |
| printk("%s: Downloading Sequencer Program...", |
| ahd_name(ahd)); |
| |
| #if DOWNLOAD_CONST_COUNT != 8 |
| #error "Download Const Mismatch" |
| #endif |
| /* |
| * Start out with 0 critical sections |
| * that apply to this firmware load. |
| */ |
| cs_count = 0; |
| cur_cs = 0; |
| memset(begin_set, 0, sizeof(begin_set)); |
| memset(end_set, 0, sizeof(end_set)); |
| |
| /* |
| * Setup downloadable constant table. |
| * |
| * The computation for the S/G prefetch variables is |
| * a bit complicated. We would like to always fetch |
| * in terms of cachelined sized increments. However, |
| * if the cacheline is not an even multiple of the |
| * SG element size or is larger than our SG RAM, using |
| * just the cache size might leave us with only a portion |
| * of an SG element at the tail of a prefetch. If the |
| * cacheline is larger than our S/G prefetch buffer less |
| * the size of an SG element, we may round down to a cacheline |
| * that doesn't contain any or all of the S/G of interest |
| * within the bounds of our S/G ram. Provide variables to |
| * the sequencer that will allow it to handle these edge |
| * cases. |
| */ |
| /* Start by aligning to the nearest cacheline. */ |
| sg_prefetch_align = ahd->pci_cachesize; |
| if (sg_prefetch_align == 0) |
| sg_prefetch_align = 8; |
| /* Round down to the nearest power of 2. */ |
| while (powerof2(sg_prefetch_align) == 0) |
| sg_prefetch_align--; |
| |
| cacheline_mask = sg_prefetch_align - 1; |
| |
| /* |
| * If the cacheline boundary is greater than half our prefetch RAM |
| * we risk not being able to fetch even a single complete S/G |
| * segment if we align to that boundary. |
| */ |
| if (sg_prefetch_align > CCSGADDR_MAX/2) |
| sg_prefetch_align = CCSGADDR_MAX/2; |
| /* Start by fetching a single cacheline. */ |
| sg_prefetch_cnt = sg_prefetch_align; |
| /* |
| * Increment the prefetch count by cachelines until |
| * at least one S/G element will fit. |
| */ |
| sg_size = sizeof(struct ahd_dma_seg); |
| if ((ahd->flags & AHD_64BIT_ADDRESSING) != 0) |
| sg_size = sizeof(struct ahd_dma64_seg); |
| while (sg_prefetch_cnt < sg_size) |
| sg_prefetch_cnt += sg_prefetch_align; |
| /* |
| * If the cacheline is not an even multiple of |
| * the S/G size, we may only get a partial S/G when |
| * we align. Add a cacheline if this is the case. |
| */ |
| if ((sg_prefetch_align % sg_size) != 0 |
| && (sg_prefetch_cnt < CCSGADDR_MAX)) |
| sg_prefetch_cnt += sg_prefetch_align; |
| /* |
| * Lastly, compute a value that the sequencer can use |
| * to determine if the remainder of the CCSGRAM buffer |
| * has a full S/G element in it. |
| */ |
| sg_prefetch_cnt_limit = -(sg_prefetch_cnt - sg_size + 1); |
| download_consts[SG_PREFETCH_CNT] = sg_prefetch_cnt; |
| download_consts[SG_PREFETCH_CNT_LIMIT] = sg_prefetch_cnt_limit; |
| download_consts[SG_PREFETCH_ALIGN_MASK] = ~(sg_prefetch_align - 1); |
| download_consts[SG_PREFETCH_ADDR_MASK] = (sg_prefetch_align - 1); |
| download_consts[SG_SIZEOF] = sg_size; |
| download_consts[PKT_OVERRUN_BUFOFFSET] = |
| (ahd->overrun_buf - (uint8_t *)ahd->qoutfifo) / 256; |
| download_consts[SCB_TRANSFER_SIZE] = SCB_TRANSFER_SIZE_1BYTE_LUN; |
| download_consts[CACHELINE_MASK] = cacheline_mask; |
| cur_patch = patches; |
| downloaded = 0; |
| skip_addr = 0; |
| ahd_outb(ahd, SEQCTL0, PERRORDIS|FAILDIS|FASTMODE|LOADRAM); |
| ahd_outw(ahd, PRGMCNT, 0); |
| |
| for (i = 0; i < sizeof(seqprog)/4; i++) { |
| if (ahd_check_patch(ahd, &cur_patch, i, &skip_addr) == 0) { |
| /* |
| * Don't download this instruction as it |
| * is in a patch that was removed. |
| */ |
| continue; |
| } |
| /* |
| * Move through the CS table until we find a CS |
| * that might apply to this instruction. |
| */ |
| for (; cur_cs < num_critical_sections; cur_cs++) { |
| if (critical_sections[cur_cs].end <= i) { |
| if (begin_set[cs_count] == TRUE |
| && end_set[cs_count] == FALSE) { |
| cs_table[cs_count].end = downloaded; |
| end_set[cs_count] = TRUE; |
| cs_count++; |
| } |
| continue; |
| } |
| if (critical_sections[cur_cs].begin <= i |
| && begin_set[cs_count] == FALSE) { |
| cs_table[cs_count].begin = downloaded; |
| begin_set[cs_count] = TRUE; |
| } |
| break; |
| } |
| ahd_download_instr(ahd, i, download_consts); |
| downloaded++; |
| } |
| |
| ahd->num_critical_sections = cs_count; |
| if (cs_count != 0) { |
| |
| cs_count *= sizeof(struct cs); |
| ahd->critical_sections = kmalloc(cs_count, GFP_ATOMIC); |
| if (ahd->critical_sections == NULL) |
| panic("ahd_loadseq: Could not malloc"); |
| memcpy(ahd->critical_sections, cs_table, cs_count); |
| } |
| ahd_outb(ahd, SEQCTL0, PERRORDIS|FAILDIS|FASTMODE); |
| |
| if (bootverbose) { |
| printk(" %d instructions downloaded\n", downloaded); |
| printk("%s: Features 0x%x, Bugs 0x%x, Flags 0x%x\n", |
| ahd_name(ahd), ahd->features, ahd->bugs, ahd->flags); |
| } |
| } |
| |
| static int |
| ahd_check_patch(struct ahd_softc *ahd, const struct patch **start_patch, |
| u_int start_instr, u_int *skip_addr) |
| { |
| const struct patch *cur_patch; |
| const struct patch *last_patch; |
| u_int num_patches; |
| |
| num_patches = ARRAY_SIZE(patches); |
| last_patch = &patches[num_patches]; |
| cur_patch = *start_patch; |
| |
| while (cur_patch < last_patch && start_instr == cur_patch->begin) { |
| |
| if (cur_patch->patch_func(ahd) == 0) { |
| |
| /* Start rejecting code */ |
| *skip_addr = start_instr + cur_patch->skip_instr; |
| cur_patch += cur_patch->skip_patch; |
| } else { |
| /* Accepted this patch. Advance to the next |
| * one and wait for our intruction pointer to |
| * hit this point. |
| */ |
| cur_patch++; |
| } |
| } |
| |
| *start_patch = cur_patch; |
| if (start_instr < *skip_addr) |
| /* Still skipping */ |
| return (0); |
| |
| return (1); |
| } |
| |
| static u_int |
| ahd_resolve_seqaddr(struct ahd_softc *ahd, u_int address) |
| { |
| const struct patch *cur_patch; |
| int address_offset; |
| u_int skip_addr; |
| u_int i; |
| |
| address_offset = 0; |
| cur_patch = patches; |
| skip_addr = 0; |
| |
| for (i = 0; i < address;) { |
| |
| ahd_check_patch(ahd, &cur_patch, i, &skip_addr); |
| |
| if (skip_addr > i) { |
| int end_addr; |
| |
| end_addr = min(address, skip_addr); |
| address_offset += end_addr - i; |
| i = skip_addr; |
| } else { |
| i++; |
| } |
| } |
| return (address - address_offset); |
| } |
| |
| static void |
| ahd_download_instr(struct ahd_softc *ahd, u_int instrptr, uint8_t *dconsts) |
| { |
| union ins_formats instr; |
| struct ins_format1 *fmt1_ins; |
| struct ins_format3 *fmt3_ins; |
| u_int opcode; |
| |
| /* |
| * The firmware is always compiled into a little endian format. |
| */ |
| instr.integer = ahd_le32toh(*(uint32_t*)&seqprog[instrptr * 4]); |
| |
| fmt1_ins = &instr.format1; |
| fmt3_ins = NULL; |
| |
| /* Pull the opcode */ |
| opcode = instr.format1.opcode; |
| switch (opcode) { |
| case AIC_OP_JMP: |
| case AIC_OP_JC: |
| case AIC_OP_JNC: |
| case AIC_OP_CALL: |
| case AIC_OP_JNE: |
| case AIC_OP_JNZ: |
| case AIC_OP_JE: |
| case AIC_OP_JZ: |
| { |
| fmt3_ins = &instr.format3; |
| fmt3_ins->address = ahd_resolve_seqaddr(ahd, fmt3_ins->address); |
| /* FALLTHROUGH */ |
| } |
| case AIC_OP_OR: |
| case AIC_OP_AND: |
| case AIC_OP_XOR: |
| case AIC_OP_ADD: |
| case AIC_OP_ADC: |
| case AIC_OP_BMOV: |
| if (fmt1_ins->parity != 0) { |
| fmt1_ins->immediate = dconsts[fmt1_ins->immediate]; |
| } |
| fmt1_ins->parity = 0; |
| /* FALLTHROUGH */ |
| case AIC_OP_ROL: |
| { |
| int i, count; |
| |
| /* Calculate odd parity for the instruction */ |
| for (i = 0, count = 0; i < 31; i++) { |
| uint32_t mask; |
| |
| mask = 0x01 << i; |
| if ((instr.integer & mask) != 0) |
| count++; |
| } |
| if ((count & 0x01) == 0) |
| instr.format1.parity = 1; |
| |
| /* The sequencer is a little endian cpu */ |
| instr.integer = ahd_htole32(instr.integer); |
| ahd_outsb(ahd, SEQRAM, instr.bytes, 4); |
| break; |
| } |
| default: |
| panic("Unknown opcode encountered in seq program"); |
| break; |
| } |
| } |
| |
| static int |
| ahd_probe_stack_size(struct ahd_softc *ahd) |
| { |
| int last_probe; |
| |
| last_probe = 0; |
| while (1) { |
| int i; |
| |
| /* |
| * We avoid using 0 as a pattern to avoid |
| * confusion if the stack implementation |
| * "back-fills" with zeros when "poping' |
| * entries. |
| */ |
| for (i = 1; i <= last_probe+1; i++) { |
| ahd_outb(ahd, STACK, i & 0xFF); |
| ahd_outb(ahd, STACK, (i >> 8) & 0xFF); |
| } |
| |
| /* Verify */ |
| for (i = last_probe+1; i > 0; i--) { |
| u_int stack_entry; |
| |
| stack_entry = ahd_inb(ahd, STACK) |
| |(ahd_inb(ahd, STACK) << 8); |
| if (stack_entry != i) |
| goto sized; |
| } |
| last_probe++; |
| } |
| sized: |
| return (last_probe); |
| } |
| |
| int |
| ahd_print_register(const ahd_reg_parse_entry_t *table, u_int num_entries, |
| const char *name, u_int address, u_int value, |
| u_int *cur_column, u_int wrap_point) |
| { |
| int printed; |
| u_int printed_mask; |
| |
| if (cur_column != NULL && *cur_column >= wrap_point) { |
| printk("\n"); |
| *cur_column = 0; |
| } |
| printed = printk("%s[0x%x]", name, value); |
| if (table == NULL) { |
| printed += printk(" "); |
| *cur_column += printed; |
| return (printed); |
| } |
| printed_mask = 0; |
| while (printed_mask != 0xFF) { |
| int entry; |
| |
| for (entry = 0; entry < num_entries; entry++) { |
| if (((value & table[entry].mask) |
| != table[entry].value) |
| || ((printed_mask & table[entry].mask) |
| == table[entry].mask)) |
| continue; |
| |
| printed += printk("%s%s", |
| printed_mask == 0 ? ":(" : "|", |
| table[entry].name); |
| printed_mask |= table[entry].mask; |
| |
| break; |
| } |
| if (entry >= num_entries) |
| break; |
| } |
| if (printed_mask != 0) |
| printed += printk(") "); |
| else |
| printed += printk(" "); |
| if (cur_column != NULL) |
| *cur_column += printed; |
| return (printed); |
| } |
| |
| void |
| ahd_dump_card_state(struct ahd_softc *ahd) |
| { |
| struct scb *scb; |
| ahd_mode_state saved_modes; |
| u_int dffstat; |
| int paused; |
| u_int scb_index; |
| u_int saved_scb_index; |
| u_int cur_col; |
| int i; |
| |
| if (ahd_is_paused(ahd)) { |
| paused = 1; |
| } else { |
| paused = 0; |
| ahd_pause(ahd); |
| } |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| printk(">>>>>>>>>>>>>>>>>> Dump Card State Begins <<<<<<<<<<<<<<<<<\n" |
| "%s: Dumping Card State at program address 0x%x Mode 0x%x\n", |
| ahd_name(ahd), |
| ahd_inw(ahd, CURADDR), |
| ahd_build_mode_state(ahd, ahd->saved_src_mode, |
| ahd->saved_dst_mode)); |
| if (paused) |
| printk("Card was paused\n"); |
| |
| if (ahd_check_cmdcmpltqueues(ahd)) |
| printk("Completions are pending\n"); |
| |
| /* |
| * Mode independent registers. |
| */ |
| cur_col = 0; |
| ahd_intstat_print(ahd_inb(ahd, INTSTAT), &cur_col, 50); |
| ahd_seloid_print(ahd_inb(ahd, SELOID), &cur_col, 50); |
| ahd_selid_print(ahd_inb(ahd, SELID), &cur_col, 50); |
| ahd_hs_mailbox_print(ahd_inb(ahd, LOCAL_HS_MAILBOX), &cur_col, 50); |
| ahd_intctl_print(ahd_inb(ahd, INTCTL), &cur_col, 50); |
| ahd_seqintstat_print(ahd_inb(ahd, SEQINTSTAT), &cur_col, 50); |
| ahd_saved_mode_print(ahd_inb(ahd, SAVED_MODE), &cur_col, 50); |
| ahd_dffstat_print(ahd_inb(ahd, DFFSTAT), &cur_col, 50); |
| ahd_scsisigi_print(ahd_inb(ahd, SCSISIGI), &cur_col, 50); |
| ahd_scsiphase_print(ahd_inb(ahd, SCSIPHASE), &cur_col, 50); |
| ahd_scsibus_print(ahd_inb(ahd, SCSIBUS), &cur_col, 50); |
| ahd_lastphase_print(ahd_inb(ahd, LASTPHASE), &cur_col, 50); |
| ahd_scsiseq0_print(ahd_inb(ahd, SCSISEQ0), &cur_col, 50); |
| ahd_scsiseq1_print(ahd_inb(ahd, SCSISEQ1), &cur_col, 50); |
| ahd_seqctl0_print(ahd_inb(ahd, SEQCTL0), &cur_col, 50); |
| ahd_seqintctl_print(ahd_inb(ahd, SEQINTCTL), &cur_col, 50); |
| ahd_seq_flags_print(ahd_inb(ahd, SEQ_FLAGS), &cur_col, 50); |
| ahd_seq_flags2_print(ahd_inb(ahd, SEQ_FLAGS2), &cur_col, 50); |
| ahd_qfreeze_count_print(ahd_inw(ahd, QFREEZE_COUNT), &cur_col, 50); |
| ahd_kernel_qfreeze_count_print(ahd_inw(ahd, KERNEL_QFREEZE_COUNT), |
| &cur_col, 50); |
| ahd_mk_message_scb_print(ahd_inw(ahd, MK_MESSAGE_SCB), &cur_col, 50); |
| ahd_mk_message_scsiid_print(ahd_inb(ahd, MK_MESSAGE_SCSIID), |
| &cur_col, 50); |
| ahd_sstat0_print(ahd_inb(ahd, SSTAT0), &cur_col, 50); |
| ahd_sstat1_print(ahd_inb(ahd, SSTAT1), &cur_col, 50); |
| ahd_sstat2_print(ahd_inb(ahd, SSTAT2), &cur_col, 50); |
| ahd_sstat3_print(ahd_inb(ahd, SSTAT3), &cur_col, 50); |
| ahd_perrdiag_print(ahd_inb(ahd, PERRDIAG), &cur_col, 50); |
| ahd_simode1_print(ahd_inb(ahd, SIMODE1), &cur_col, 50); |
| ahd_lqistat0_print(ahd_inb(ahd, LQISTAT0), &cur_col, 50); |
| ahd_lqistat1_print(ahd_inb(ahd, LQISTAT1), &cur_col, 50); |
| ahd_lqistat2_print(ahd_inb(ahd, LQISTAT2), &cur_col, 50); |
| ahd_lqostat0_print(ahd_inb(ahd, LQOSTAT0), &cur_col, 50); |
| ahd_lqostat1_print(ahd_inb(ahd, LQOSTAT1), &cur_col, 50); |
| ahd_lqostat2_print(ahd_inb(ahd, LQOSTAT2), &cur_col, 50); |
| printk("\n"); |
| printk("\nSCB Count = %d CMDS_PENDING = %d LASTSCB 0x%x " |
| "CURRSCB 0x%x NEXTSCB 0x%x\n", |
| ahd->scb_data.numscbs, ahd_inw(ahd, CMDS_PENDING), |
| ahd_inw(ahd, LASTSCB), ahd_inw(ahd, CURRSCB), |
| ahd_inw(ahd, NEXTSCB)); |
| cur_col = 0; |
| /* QINFIFO */ |
| ahd_search_qinfifo(ahd, CAM_TARGET_WILDCARD, ALL_CHANNELS, |
| CAM_LUN_WILDCARD, SCB_LIST_NULL, |
| ROLE_UNKNOWN, /*status*/0, SEARCH_PRINT); |
| saved_scb_index = ahd_get_scbptr(ahd); |
| printk("Pending list:"); |
| i = 0; |
| LIST_FOREACH(scb, &ahd->pending_scbs, pending_links) { |
| if (i++ > AHD_SCB_MAX) |
| break; |
| cur_col = printk("\n%3d FIFO_USE[0x%x] ", SCB_GET_TAG(scb), |
| ahd_inb_scbram(ahd, SCB_FIFO_USE_COUNT)); |
| ahd_set_scbptr(ahd, SCB_GET_TAG(scb)); |
| ahd_scb_control_print(ahd_inb_scbram(ahd, SCB_CONTROL), |
| &cur_col, 60); |
| ahd_scb_scsiid_print(ahd_inb_scbram(ahd, SCB_SCSIID), |
| &cur_col, 60); |
| } |
| printk("\nTotal %d\n", i); |
| |
| printk("Kernel Free SCB list: "); |
| i = 0; |
| TAILQ_FOREACH(scb, &ahd->scb_data.free_scbs, links.tqe) { |
| struct scb *list_scb; |
| |
| list_scb = scb; |
| do { |
| printk("%d ", SCB_GET_TAG(list_scb)); |
| list_scb = LIST_NEXT(list_scb, collision_links); |
| } while (list_scb && i++ < AHD_SCB_MAX); |
| } |
| |
| LIST_FOREACH(scb, &ahd->scb_data.any_dev_free_scb_list, links.le) { |
| if (i++ > AHD_SCB_MAX) |
| break; |
| printk("%d ", SCB_GET_TAG(scb)); |
| } |
| printk("\n"); |
| |
| printk("Sequencer Complete DMA-inprog list: "); |
| scb_index = ahd_inw(ahd, COMPLETE_SCB_DMAINPROG_HEAD); |
| i = 0; |
| while (!SCBID_IS_NULL(scb_index) && i++ < AHD_SCB_MAX) { |
| ahd_set_scbptr(ahd, scb_index); |
| printk("%d ", scb_index); |
| scb_index = ahd_inw_scbram(ahd, SCB_NEXT_COMPLETE); |
| } |
| printk("\n"); |
| |
| printk("Sequencer Complete list: "); |
| scb_index = ahd_inw(ahd, COMPLETE_SCB_HEAD); |
| i = 0; |
| while (!SCBID_IS_NULL(scb_index) && i++ < AHD_SCB_MAX) { |
| ahd_set_scbptr(ahd, scb_index); |
| printk("%d ", scb_index); |
| scb_index = ahd_inw_scbram(ahd, SCB_NEXT_COMPLETE); |
| } |
| printk("\n"); |
| |
| |
| printk("Sequencer DMA-Up and Complete list: "); |
| scb_index = ahd_inw(ahd, COMPLETE_DMA_SCB_HEAD); |
| i = 0; |
| while (!SCBID_IS_NULL(scb_index) && i++ < AHD_SCB_MAX) { |
| ahd_set_scbptr(ahd, scb_index); |
| printk("%d ", scb_index); |
| scb_index = ahd_inw_scbram(ahd, SCB_NEXT_COMPLETE); |
| } |
| printk("\n"); |
| printk("Sequencer On QFreeze and Complete list: "); |
| scb_index = ahd_inw(ahd, COMPLETE_ON_QFREEZE_HEAD); |
| i = 0; |
| while (!SCBID_IS_NULL(scb_index) && i++ < AHD_SCB_MAX) { |
| ahd_set_scbptr(ahd, scb_index); |
| printk("%d ", scb_index); |
| scb_index = ahd_inw_scbram(ahd, SCB_NEXT_COMPLETE); |
| } |
| printk("\n"); |
| ahd_set_scbptr(ahd, saved_scb_index); |
| dffstat = ahd_inb(ahd, DFFSTAT); |
| for (i = 0; i < 2; i++) { |
| #ifdef AHD_DEBUG |
| struct scb *fifo_scb; |
| #endif |
| u_int fifo_scbptr; |
| |
| ahd_set_modes(ahd, AHD_MODE_DFF0 + i, AHD_MODE_DFF0 + i); |
| fifo_scbptr = ahd_get_scbptr(ahd); |
| printk("\n\n%s: FIFO%d %s, LONGJMP == 0x%x, SCB 0x%x\n", |
| ahd_name(ahd), i, |
| (dffstat & (FIFO0FREE << i)) ? "Free" : "Active", |
| ahd_inw(ahd, LONGJMP_ADDR), fifo_scbptr); |
| cur_col = 0; |
| ahd_seqimode_print(ahd_inb(ahd, SEQIMODE), &cur_col, 50); |
| ahd_seqintsrc_print(ahd_inb(ahd, SEQINTSRC), &cur_col, 50); |
| ahd_dfcntrl_print(ahd_inb(ahd, DFCNTRL), &cur_col, 50); |
| ahd_dfstatus_print(ahd_inb(ahd, DFSTATUS), &cur_col, 50); |
| ahd_sg_cache_shadow_print(ahd_inb(ahd, SG_CACHE_SHADOW), |
| &cur_col, 50); |
| ahd_sg_state_print(ahd_inb(ahd, SG_STATE), &cur_col, 50); |
| ahd_dffsxfrctl_print(ahd_inb(ahd, DFFSXFRCTL), &cur_col, 50); |
| ahd_soffcnt_print(ahd_inb(ahd, SOFFCNT), &cur_col, 50); |
| ahd_mdffstat_print(ahd_inb(ahd, MDFFSTAT), &cur_col, 50); |
| if (cur_col > 50) { |
| printk("\n"); |
| cur_col = 0; |
| } |
| cur_col += printk("SHADDR = 0x%x%x, SHCNT = 0x%x ", |
| ahd_inl(ahd, SHADDR+4), |
| ahd_inl(ahd, SHADDR), |
| (ahd_inb(ahd, SHCNT) |
| | (ahd_inb(ahd, SHCNT + 1) << 8) |
| | (ahd_inb(ahd, SHCNT + 2) << 16))); |
| if (cur_col > 50) { |
| printk("\n"); |
| cur_col = 0; |
| } |
| cur_col += printk("HADDR = 0x%x%x, HCNT = 0x%x ", |
| ahd_inl(ahd, HADDR+4), |
| ahd_inl(ahd, HADDR), |
| (ahd_inb(ahd, HCNT) |
| | (ahd_inb(ahd, HCNT + 1) << 8) |
| | (ahd_inb(ahd, HCNT + 2) << 16))); |
| ahd_ccsgctl_print(ahd_inb(ahd, CCSGCTL), &cur_col, 50); |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_SG) != 0) { |
| fifo_scb = ahd_lookup_scb(ahd, fifo_scbptr); |
| if (fifo_scb != NULL) |
| ahd_dump_sglist(fifo_scb); |
| } |
| #endif |
| } |
| printk("\nLQIN: "); |
| for (i = 0; i < 20; i++) |
| printk("0x%x ", ahd_inb(ahd, LQIN + i)); |
| printk("\n"); |
| ahd_set_modes(ahd, AHD_MODE_CFG, AHD_MODE_CFG); |
| printk("%s: LQISTATE = 0x%x, LQOSTATE = 0x%x, OPTIONMODE = 0x%x\n", |
| ahd_name(ahd), ahd_inb(ahd, LQISTATE), ahd_inb(ahd, LQOSTATE), |
| ahd_inb(ahd, OPTIONMODE)); |
| printk("%s: OS_SPACE_CNT = 0x%x MAXCMDCNT = 0x%x\n", |
| ahd_name(ahd), ahd_inb(ahd, OS_SPACE_CNT), |
| ahd_inb(ahd, MAXCMDCNT)); |
| printk("%s: SAVED_SCSIID = 0x%x SAVED_LUN = 0x%x\n", |
| ahd_name(ahd), ahd_inb(ahd, SAVED_SCSIID), |
| ahd_inb(ahd, SAVED_LUN)); |
| ahd_simode0_print(ahd_inb(ahd, SIMODE0), &cur_col, 50); |
| printk("\n"); |
| ahd_set_modes(ahd, AHD_MODE_CCHAN, AHD_MODE_CCHAN); |
| cur_col = 0; |
| ahd_ccscbctl_print(ahd_inb(ahd, CCSCBCTL), &cur_col, 50); |
| printk("\n"); |
| ahd_set_modes(ahd, ahd->saved_src_mode, ahd->saved_dst_mode); |
| printk("%s: REG0 == 0x%x, SINDEX = 0x%x, DINDEX = 0x%x\n", |
| ahd_name(ahd), ahd_inw(ahd, REG0), ahd_inw(ahd, SINDEX), |
| ahd_inw(ahd, DINDEX)); |
| printk("%s: SCBPTR == 0x%x, SCB_NEXT == 0x%x, SCB_NEXT2 == 0x%x\n", |
| ahd_name(ahd), ahd_get_scbptr(ahd), |
| ahd_inw_scbram(ahd, SCB_NEXT), |
| ahd_inw_scbram(ahd, SCB_NEXT2)); |
| printk("CDB %x %x %x %x %x %x\n", |
| ahd_inb_scbram(ahd, SCB_CDB_STORE), |
| ahd_inb_scbram(ahd, SCB_CDB_STORE+1), |
| ahd_inb_scbram(ahd, SCB_CDB_STORE+2), |
| ahd_inb_scbram(ahd, SCB_CDB_STORE+3), |
| ahd_inb_scbram(ahd, SCB_CDB_STORE+4), |
| ahd_inb_scbram(ahd, SCB_CDB_STORE+5)); |
| printk("STACK:"); |
| for (i = 0; i < ahd->stack_size; i++) { |
| ahd->saved_stack[i] = |
| ahd_inb(ahd, STACK)|(ahd_inb(ahd, STACK) << 8); |
| printk(" 0x%x", ahd->saved_stack[i]); |
| } |
| for (i = ahd->stack_size-1; i >= 0; i--) { |
| ahd_outb(ahd, STACK, ahd->saved_stack[i] & 0xFF); |
| ahd_outb(ahd, STACK, (ahd->saved_stack[i] >> 8) & 0xFF); |
| } |
| printk("\n<<<<<<<<<<<<<<<<< Dump Card State Ends >>>>>>>>>>>>>>>>>>\n"); |
| ahd_restore_modes(ahd, saved_modes); |
| if (paused == 0) |
| ahd_unpause(ahd); |
| } |
| |
| #if 0 |
| void |
| ahd_dump_scbs(struct ahd_softc *ahd) |
| { |
| ahd_mode_state saved_modes; |
| u_int saved_scb_index; |
| int i; |
| |
| saved_modes = ahd_save_modes(ahd); |
| ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI); |
| saved_scb_index = ahd_get_scbptr(ahd); |
| for (i = 0; i < AHD_SCB_MAX; i++) { |
| ahd_set_scbptr(ahd, i); |
| printk("%3d", i); |
| printk("(CTRL 0x%x ID 0x%x N 0x%x N2 0x%x SG 0x%x, RSG 0x%x)\n", |
| ahd_inb_scbram(ahd, SCB_CONTROL), |
| ahd_inb_scbram(ahd, SCB_SCSIID), |
| ahd_inw_scbram(ahd, SCB_NEXT), |
| ahd_inw_scbram(ahd, SCB_NEXT2), |
| ahd_inl_scbram(ahd, SCB_SGPTR), |
| ahd_inl_scbram(ahd, SCB_RESIDUAL_SGPTR)); |
| } |
| printk("\n"); |
| ahd_set_scbptr(ahd, saved_scb_index); |
| ahd_restore_modes(ahd, saved_modes); |
| } |
| #endif /* 0 */ |
| |
| /**************************** Flexport Logic **********************************/ |
| /* |
| * Read count 16bit words from 16bit word address start_addr from the |
| * SEEPROM attached to the controller, into buf, using the controller's |
| * SEEPROM reading state machine. Optionally treat the data as a byte |
| * stream in terms of byte order. |
| */ |
| int |
| ahd_read_seeprom(struct ahd_softc *ahd, uint16_t *buf, |
| u_int start_addr, u_int count, int bytestream) |
| { |
| u_int cur_addr; |
| u_int end_addr; |
| int error; |
| |
| /* |
| * If we never make it through the loop even once, |
| * we were passed invalid arguments. |
| */ |
| error = EINVAL; |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| end_addr = start_addr + count; |
| for (cur_addr = start_addr; cur_addr < end_addr; cur_addr++) { |
| |
| ahd_outb(ahd, SEEADR, cur_addr); |
| ahd_outb(ahd, SEECTL, SEEOP_READ | SEESTART); |
| |
| error = ahd_wait_seeprom(ahd); |
| if (error) |
| break; |
| if (bytestream != 0) { |
| uint8_t *bytestream_ptr; |
| |
| bytestream_ptr = (uint8_t *)buf; |
| *bytestream_ptr++ = ahd_inb(ahd, SEEDAT); |
| *bytestream_ptr = ahd_inb(ahd, SEEDAT+1); |
| } else { |
| /* |
| * ahd_inw() already handles machine byte order. |
| */ |
| *buf = ahd_inw(ahd, SEEDAT); |
| } |
| buf++; |
| } |
| return (error); |
| } |
| |
| /* |
| * Write count 16bit words from buf, into SEEPROM attache to the |
| * controller starting at 16bit word address start_addr, using the |
| * controller's SEEPROM writing state machine. |
| */ |
| int |
| ahd_write_seeprom(struct ahd_softc *ahd, uint16_t *buf, |
| u_int start_addr, u_int count) |
| { |
| u_int cur_addr; |
| u_int end_addr; |
| int error; |
| int retval; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| error = ENOENT; |
| |
| /* Place the chip into write-enable mode */ |
| ahd_outb(ahd, SEEADR, SEEOP_EWEN_ADDR); |
| ahd_outb(ahd, SEECTL, SEEOP_EWEN | SEESTART); |
| error = ahd_wait_seeprom(ahd); |
| if (error) |
| return (error); |
| |
| /* |
| * Write the data. If we don't get through the loop at |
| * least once, the arguments were invalid. |
| */ |
| retval = EINVAL; |
| end_addr = start_addr + count; |
| for (cur_addr = start_addr; cur_addr < end_addr; cur_addr++) { |
| ahd_outw(ahd, SEEDAT, *buf++); |
| ahd_outb(ahd, SEEADR, cur_addr); |
| ahd_outb(ahd, SEECTL, SEEOP_WRITE | SEESTART); |
| |
| retval = ahd_wait_seeprom(ahd); |
| if (retval) |
| break; |
| } |
| |
| /* |
| * Disable writes. |
| */ |
| ahd_outb(ahd, SEEADR, SEEOP_EWDS_ADDR); |
| ahd_outb(ahd, SEECTL, SEEOP_EWDS | SEESTART); |
| error = ahd_wait_seeprom(ahd); |
| if (error) |
| return (error); |
| return (retval); |
| } |
| |
| /* |
| * Wait ~100us for the serial eeprom to satisfy our request. |
| */ |
| static int |
| ahd_wait_seeprom(struct ahd_softc *ahd) |
| { |
| int cnt; |
| |
| cnt = 5000; |
| while ((ahd_inb(ahd, SEESTAT) & (SEEARBACK|SEEBUSY)) != 0 && --cnt) |
| ahd_delay(5); |
| |
| if (cnt == 0) |
| return (ETIMEDOUT); |
| return (0); |
| } |
| |
| /* |
| * Validate the two checksums in the per_channel |
| * vital product data struct. |
| */ |
| static int |
| ahd_verify_vpd_cksum(struct vpd_config *vpd) |
| { |
| int i; |
| int maxaddr; |
| uint32_t checksum; |
| uint8_t *vpdarray; |
| |
| vpdarray = (uint8_t *)vpd; |
| maxaddr = offsetof(struct vpd_config, vpd_checksum); |
| checksum = 0; |
| for (i = offsetof(struct vpd_config, resource_type); i < maxaddr; i++) |
| checksum = checksum + vpdarray[i]; |
| if (checksum == 0 |
| || (-checksum & 0xFF) != vpd->vpd_checksum) |
| return (0); |
| |
| checksum = 0; |
| maxaddr = offsetof(struct vpd_config, checksum); |
| for (i = offsetof(struct vpd_config, default_target_flags); |
| i < maxaddr; i++) |
| checksum = checksum + vpdarray[i]; |
| if (checksum == 0 |
| || (-checksum & 0xFF) != vpd->checksum) |
| return (0); |
| return (1); |
| } |
| |
| int |
| ahd_verify_cksum(struct seeprom_config *sc) |
| { |
| int i; |
| int maxaddr; |
| uint32_t checksum; |
| uint16_t *scarray; |
| |
| maxaddr = (sizeof(*sc)/2) - 1; |
| checksum = 0; |
| scarray = (uint16_t *)sc; |
| |
| for (i = 0; i < maxaddr; i++) |
| checksum = checksum + scarray[i]; |
| if (checksum == 0 |
| || (checksum & 0xFFFF) != sc->checksum) { |
| return (0); |
| } else { |
| return (1); |
| } |
| } |
| |
| int |
| ahd_acquire_seeprom(struct ahd_softc *ahd) |
| { |
| /* |
| * We should be able to determine the SEEPROM type |
| * from the flexport logic, but unfortunately not |
| * all implementations have this logic and there is |
| * no programatic method for determining if the logic |
| * is present. |
| */ |
| return (1); |
| #if 0 |
| uint8_t seetype; |
| int error; |
| |
| error = ahd_read_flexport(ahd, FLXADDR_ROMSTAT_CURSENSECTL, &seetype); |
| if (error != 0 |
| || ((seetype & FLX_ROMSTAT_SEECFG) == FLX_ROMSTAT_SEE_NONE)) |
| return (0); |
| return (1); |
| #endif |
| } |
| |
| void |
| ahd_release_seeprom(struct ahd_softc *ahd) |
| { |
| /* Currently a no-op */ |
| } |
| |
| /* |
| * Wait at most 2 seconds for flexport arbitration to succeed. |
| */ |
| static int |
| ahd_wait_flexport(struct ahd_softc *ahd) |
| { |
| int cnt; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| cnt = 1000000 * 2 / 5; |
| while ((ahd_inb(ahd, BRDCTL) & FLXARBACK) == 0 && --cnt) |
| ahd_delay(5); |
| |
| if (cnt == 0) |
| return (ETIMEDOUT); |
| return (0); |
| } |
| |
| int |
| ahd_write_flexport(struct ahd_softc *ahd, u_int addr, u_int value) |
| { |
| int error; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| if (addr > 7) |
| panic("ahd_write_flexport: address out of range"); |
| ahd_outb(ahd, BRDCTL, BRDEN|(addr << 3)); |
| error = ahd_wait_flexport(ahd); |
| if (error != 0) |
| return (error); |
| ahd_outb(ahd, BRDDAT, value); |
| ahd_flush_device_writes(ahd); |
| ahd_outb(ahd, BRDCTL, BRDSTB|BRDEN|(addr << 3)); |
| ahd_flush_device_writes(ahd); |
| ahd_outb(ahd, BRDCTL, BRDEN|(addr << 3)); |
| ahd_flush_device_writes(ahd); |
| ahd_outb(ahd, BRDCTL, 0); |
| ahd_flush_device_writes(ahd); |
| return (0); |
| } |
| |
| int |
| ahd_read_flexport(struct ahd_softc *ahd, u_int addr, uint8_t *value) |
| { |
| int error; |
| |
| AHD_ASSERT_MODES(ahd, AHD_MODE_SCSI_MSK, AHD_MODE_SCSI_MSK); |
| if (addr > 7) |
| panic("ahd_read_flexport: address out of range"); |
| ahd_outb(ahd, BRDCTL, BRDRW|BRDEN|(addr << 3)); |
| error = ahd_wait_flexport(ahd); |
| if (error != 0) |
| return (error); |
| *value = ahd_inb(ahd, BRDDAT); |
| ahd_outb(ahd, BRDCTL, 0); |
| ahd_flush_device_writes(ahd); |
| return (0); |
| } |
| |
| /************************* Target Mode ****************************************/ |
| #ifdef AHD_TARGET_MODE |
| cam_status |
| ahd_find_tmode_devs(struct ahd_softc *ahd, struct cam_sim *sim, union ccb *ccb, |
| struct ahd_tmode_tstate **tstate, |
| struct ahd_tmode_lstate **lstate, |
| int notfound_failure) |
| { |
| |
| if ((ahd->features & AHD_TARGETMODE) == 0) |
| return (CAM_REQ_INVALID); |
| |
| /* |
| * Handle the 'black hole' device that sucks up |
| * requests to unattached luns on enabled targets. |
| */ |
| if (ccb->ccb_h.target_id == CAM_TARGET_WILDCARD |
| && ccb->ccb_h.target_lun == CAM_LUN_WILDCARD) { |
| *tstate = NULL; |
| *lstate = ahd->black_hole; |
| } else { |
| u_int max_id; |
| |
| max_id = (ahd->features & AHD_WIDE) ? 16 : 8; |
| if (ccb->ccb_h.target_id >= max_id) |
| return (CAM_TID_INVALID); |
| |
| if (ccb->ccb_h.target_lun >= AHD_NUM_LUNS) |
| return (CAM_LUN_INVALID); |
| |
| *tstate = ahd->enabled_targets[ccb->ccb_h.target_id]; |
| *lstate = NULL; |
| if (*tstate != NULL) |
| *lstate = |
| (*tstate)->enabled_luns[ccb->ccb_h.target_lun]; |
| } |
| |
| if (notfound_failure != 0 && *lstate == NULL) |
| return (CAM_PATH_INVALID); |
| |
| return (CAM_REQ_CMP); |
| } |
| |
| void |
| ahd_handle_en_lun(struct ahd_softc *ahd, struct cam_sim *sim, union ccb *ccb) |
| { |
| #if NOT_YET |
| struct ahd_tmode_tstate *tstate; |
| struct ahd_tmode_lstate *lstate; |
| struct ccb_en_lun *cel; |
| cam_status status; |
| u_int target; |
| u_int lun; |
| u_int target_mask; |
| u_long s; |
| char channel; |
| |
| status = ahd_find_tmode_devs(ahd, sim, ccb, &tstate, &lstate, |
| /*notfound_failure*/FALSE); |
| |
| if (status != CAM_REQ_CMP) { |
| ccb->ccb_h.status = status; |
| return; |
| } |
| |
| if ((ahd->features & AHD_MULTIROLE) != 0) { |
| u_int our_id; |
| |
| our_id = ahd->our_id; |
| if (ccb->ccb_h.target_id != our_id) { |
| if ((ahd->features & AHD_MULTI_TID) != 0 |
| && (ahd->flags & AHD_INITIATORROLE) != 0) { |
| /* |
| * Only allow additional targets if |
| * the initiator role is disabled. |
| * The hardware cannot handle a re-select-in |
| * on the initiator id during a re-select-out |
| * on a different target id. |
| */ |
| status = CAM_TID_INVALID; |
| } else if ((ahd->flags & AHD_INITIATORROLE) != 0 |
| || ahd->enabled_luns > 0) { |
| /* |
| * Only allow our target id to change |
| * if the initiator role is not configured |
| * and there are no enabled luns which |
| * are attached to the currently registered |
| * scsi id. |
| */ |
| status = CAM_TID_INVALID; |
| } |
| } |
| } |
| |
| if (status != CAM_REQ_CMP) { |
| ccb->ccb_h.status = status; |
| return; |
| } |
| |
| /* |
| * We now have an id that is valid. |
| * If we aren't in target mode, switch modes. |
| */ |
| if ((ahd->flags & AHD_TARGETROLE) == 0 |
| && ccb->ccb_h.target_id != CAM_TARGET_WILDCARD) { |
| u_long s; |
| |
| printk("Configuring Target Mode\n"); |
| ahd_lock(ahd, &s); |
| if (LIST_FIRST(&ahd->pending_scbs) != NULL) { |
| ccb->ccb_h.status = CAM_BUSY; |
| ahd_unlock(ahd, &s); |
| return; |
| } |
| ahd->flags |= AHD_TARGETROLE; |
| if ((ahd->features & AHD_MULTIROLE) == 0) |
| ahd->flags &= ~AHD_INITIATORROLE; |
| ahd_pause(ahd); |
| ahd_loadseq(ahd); |
| ahd_restart(ahd); |
| ahd_unlock(ahd, &s); |
| } |
| cel = &ccb->cel; |
| target = ccb->ccb_h.target_id; |
| lun = ccb->ccb_h.target_lun; |
| channel = SIM_CHANNEL(ahd, sim); |
| target_mask = 0x01 << target; |
| if (channel == 'B') |
| target_mask <<= 8; |
| |
| if (cel->enable != 0) { |
| u_int scsiseq1; |
| |
| /* Are we already enabled?? */ |
| if (lstate != NULL) { |
| xpt_print_path(ccb->ccb_h.path); |
| printk("Lun already enabled\n"); |
| ccb->ccb_h.status = CAM_LUN_ALRDY_ENA; |
| return; |
| } |
| |
| if (cel->grp6_len != 0 |
| || cel->grp7_len != 0) { |
| /* |
| * Don't (yet?) support vendor |
| * specific commands. |
| */ |
| ccb->ccb_h.status = CAM_REQ_INVALID; |
| printk("Non-zero Group Codes\n"); |
| return; |
| } |
| |
| /* |
| * Seems to be okay. |
| * Setup our data structures. |
| */ |
| if (target != CAM_TARGET_WILDCARD && tstate == NULL) { |
| tstate = ahd_alloc_tstate(ahd, target, channel); |
| if (tstate == NULL) { |
| xpt_print_path(ccb->ccb_h.path); |
| printk("Couldn't allocate tstate\n"); |
| ccb->ccb_h.status = CAM_RESRC_UNAVAIL; |
| return; |
| } |
| } |
| lstate = kzalloc(sizeof(*lstate), GFP_ATOMIC); |
| if (lstate == NULL) { |
| xpt_print_path(ccb->ccb_h.path); |
| printk("Couldn't allocate lstate\n"); |
| ccb->ccb_h.status = CAM_RESRC_UNAVAIL; |
| return; |
| } |
| status = xpt_create_path(&lstate->path, /*periph*/NULL, |
| xpt_path_path_id(ccb->ccb_h.path), |
| xpt_path_target_id(ccb->ccb_h.path), |
| xpt_path_lun_id(ccb->ccb_h.path)); |
| if (status != CAM_REQ_CMP) { |
| kfree(lstate); |
| xpt_print_path(ccb->ccb_h.path); |
| printk("Couldn't allocate path\n"); |
| ccb->ccb_h.status = CAM_RESRC_UNAVAIL; |
| return; |
| } |
| SLIST_INIT(&lstate->accept_tios); |
| SLIST_INIT(&lstate->immed_notifies); |
| ahd_lock(ahd, &s); |
| ahd_pause(ahd); |
| if (target != CAM_TARGET_WILDCARD) { |
| tstate->enabled_luns[lun] = lstate; |
| ahd->enabled_luns++; |
| |
| if ((ahd->features & AHD_MULTI_TID) != 0) { |
| u_int targid_mask; |
| |
| targid_mask = ahd_inw(ahd, TARGID); |
| targid_mask |= target_mask; |
| ahd_outw(ahd, TARGID, targid_mask); |
| ahd_update_scsiid(ahd, targid_mask); |
| } else { |
| u_int our_id; |
| char channel; |
| |
| channel = SIM_CHANNEL(ahd, sim); |
| our_id = SIM_SCSI_ID(ahd, sim); |
| |
| /* |
| * This can only happen if selections |
| * are not enabled |
| */ |
| if (target != our_id) { |
| u_int sblkctl; |
| char cur_channel; |
| int swap; |
| |
| sblkctl = ahd_inb(ahd, SBLKCTL); |
| cur_channel = (sblkctl & SELBUSB) |
| ? 'B' : 'A'; |
| if ((ahd->features & AHD_TWIN) == 0) |
| cur_channel = 'A'; |
| swap = cur_channel != channel; |
| ahd->our_id = target; |
| |
| if (swap) |
| ahd_outb(ahd, SBLKCTL, |
| sblkctl ^ SELBUSB); |
| |
| ahd_outb(ahd, SCSIID, target); |
| |
| if (swap) |
| ahd_outb(ahd, SBLKCTL, sblkctl); |
| } |
| } |
| } else |
| ahd->black_hole = lstate; |
| /* Allow select-in operations */ |
| if (ahd->black_hole != NULL && ahd->enabled_luns > 0) { |
| scsiseq1 = ahd_inb(ahd, SCSISEQ_TEMPLATE); |
| scsiseq1 |= ENSELI; |
| ahd_outb(ahd, SCSISEQ_TEMPLATE, scsiseq1); |
| scsiseq1 = ahd_inb(ahd, SCSISEQ1); |
| scsiseq1 |= ENSELI; |
| ahd_outb(ahd, SCSISEQ1, scsiseq1); |
| } |
| ahd_unpause(ahd); |
| ahd_unlock(ahd, &s); |
| ccb->ccb_h.status = CAM_REQ_CMP; |
| xpt_print_path(ccb->ccb_h.path); |
| printk("Lun now enabled for target mode\n"); |
| } else { |
| struct scb *scb; |
| int i, empty; |
| |
| if (lstate == NULL) { |
| ccb->ccb_h.status = CAM_LUN_INVALID; |
| return; |
| } |
| |
| ahd_lock(ahd, &s); |
| |
| ccb->ccb_h.status = CAM_REQ_CMP; |
| LIST_FOREACH(scb, &ahd->pending_scbs, pending_links) { |
| struct ccb_hdr *ccbh; |
| |
| ccbh = &scb->io_ctx->ccb_h; |
| if (ccbh->func_code == XPT_CONT_TARGET_IO |
| && !xpt_path_comp(ccbh->path, ccb->ccb_h.path)){ |
| printk("CTIO pending\n"); |
| ccb->ccb_h.status = CAM_REQ_INVALID; |
| ahd_unlock(ahd, &s); |
| return; |
| } |
| } |
| |
| if (SLIST_FIRST(&lstate->accept_tios) != NULL) { |
| printk("ATIOs pending\n"); |
| ccb->ccb_h.status = CAM_REQ_INVALID; |
| } |
| |
| if (SLIST_FIRST(&lstate->immed_notifies) != NULL) { |
| printk("INOTs pending\n"); |
| ccb->ccb_h.status = CAM_REQ_INVALID; |
| } |
| |
| if (ccb->ccb_h.status != CAM_REQ_CMP) { |
| ahd_unlock(ahd, &s); |
| return; |
| } |
| |
| xpt_print_path(ccb->ccb_h.path); |
| printk("Target mode disabled\n"); |
| xpt_free_path(lstate->path); |
| kfree(lstate); |
| |
| ahd_pause(ahd); |
| /* Can we clean up the target too? */ |
| if (target != CAM_TARGET_WILDCARD) { |
| tstate->enabled_luns[lun] = NULL; |
| ahd->enabled_luns--; |
| for (empty = 1, i = 0; i < 8; i++) |
| if (tstate->enabled_luns[i] != NULL) { |
| empty = 0; |
| break; |
| } |
| |
| if (empty) { |
| ahd_free_tstate(ahd, target, channel, |
| /*force*/FALSE); |
| if (ahd->features & AHD_MULTI_TID) { |
| u_int targid_mask; |
| |
| targid_mask = ahd_inw(ahd, TARGID); |
| targid_mask &= ~target_mask; |
| ahd_outw(ahd, TARGID, targid_mask); |
| ahd_update_scsiid(ahd, targid_mask); |
| } |
| } |
| } else { |
| |
| ahd->black_hole = NULL; |
| |
| /* |
| * We can't allow selections without |
| * our black hole device. |
| */ |
| empty = TRUE; |
| } |
| if (ahd->enabled_luns == 0) { |
| /* Disallow select-in */ |
| u_int scsiseq1; |
| |
| scsiseq1 = ahd_inb(ahd, SCSISEQ_TEMPLATE); |
| scsiseq1 &= ~ENSELI; |
| ahd_outb(ahd, SCSISEQ_TEMPLATE, scsiseq1); |
| scsiseq1 = ahd_inb(ahd, SCSISEQ1); |
| scsiseq1 &= ~ENSELI; |
| ahd_outb(ahd, SCSISEQ1, scsiseq1); |
| |
| if ((ahd->features & AHD_MULTIROLE) == 0) { |
| printk("Configuring Initiator Mode\n"); |
| ahd->flags &= ~AHD_TARGETROLE; |
| ahd->flags |= AHD_INITIATORROLE; |
| ahd_pause(ahd); |
| ahd_loadseq(ahd); |
| ahd_restart(ahd); |
| /* |
| * Unpaused. The extra unpause |
| * that follows is harmless. |
| */ |
| } |
| } |
| ahd_unpause(ahd); |
| ahd_unlock(ahd, &s); |
| } |
| #endif |
| } |
| |
| static void |
| ahd_update_scsiid(struct ahd_softc *ahd, u_int targid_mask) |
| { |
| #if NOT_YET |
| u_int scsiid_mask; |
| u_int scsiid; |
| |
| if ((ahd->features & AHD_MULTI_TID) == 0) |
| panic("ahd_update_scsiid called on non-multitid unit\n"); |
| |
| /* |
| * Since we will rely on the TARGID mask |
| * for selection enables, ensure that OID |
| * in SCSIID is not set to some other ID |
| * that we don't want to allow selections on. |
| */ |
| if ((ahd->features & AHD_ULTRA2) != 0) |
| scsiid = ahd_inb(ahd, SCSIID_ULTRA2); |
| else |
| scsiid = ahd_inb(ahd, SCSIID); |
| scsiid_mask = 0x1 << (scsiid & OID); |
| if ((targid_mask & scsiid_mask) == 0) { |
| u_int our_id; |
| |
| /* ffs counts from 1 */ |
| our_id = ffs(targid_mask); |
| if (our_id == 0) |
| our_id = ahd->our_id; |
| else |
| our_id--; |
| scsiid &= TID; |
| scsiid |= our_id; |
| } |
| if ((ahd->features & AHD_ULTRA2) != 0) |
| ahd_outb(ahd, SCSIID_ULTRA2, scsiid); |
| else |
| ahd_outb(ahd, SCSIID, scsiid); |
| #endif |
| } |
| |
| static void |
| ahd_run_tqinfifo(struct ahd_softc *ahd, int paused) |
| { |
| struct target_cmd *cmd; |
| |
| ahd_sync_tqinfifo(ahd, BUS_DMASYNC_POSTREAD); |
| while ((cmd = &ahd->targetcmds[ahd->tqinfifonext])->cmd_valid != 0) { |
| |
| /* |
| * Only advance through the queue if we |
| * have the resources to process the command. |
| */ |
| if (ahd_handle_target_cmd(ahd, cmd) != 0) |
| break; |
| |
| cmd->cmd_valid = 0; |
| ahd_dmamap_sync(ahd, ahd->shared_data_dmat, |
| ahd->shared_data_map.dmamap, |
| ahd_targetcmd_offset(ahd, ahd->tqinfifonext), |
| sizeof(struct target_cmd), |
| BUS_DMASYNC_PREREAD); |
| ahd->tqinfifonext++; |
| |
| /* |
| * Lazily update our position in the target mode incoming |
| * command queue as seen by the sequencer. |
| */ |
| if ((ahd->tqinfifonext & (HOST_TQINPOS - 1)) == 1) { |
| u_int hs_mailbox; |
| |
| hs_mailbox = ahd_inb(ahd, HS_MAILBOX); |
| hs_mailbox &= ~HOST_TQINPOS; |
| hs_mailbox |= ahd->tqinfifonext & HOST_TQINPOS; |
| ahd_outb(ahd, HS_MAILBOX, hs_mailbox); |
| } |
| } |
| } |
| |
| static int |
| ahd_handle_target_cmd(struct ahd_softc *ahd, struct target_cmd *cmd) |
| { |
| struct ahd_tmode_tstate *tstate; |
| struct ahd_tmode_lstate *lstate; |
| struct ccb_accept_tio *atio; |
| uint8_t *byte; |
| int initiator; |
| int target; |
| int lun; |
| |
| initiator = SCSIID_TARGET(ahd, cmd->scsiid); |
| target = SCSIID_OUR_ID(cmd->scsiid); |
| lun = (cmd->identify & MSG_IDENTIFY_LUNMASK); |
| |
| byte = cmd->bytes; |
| tstate = ahd->enabled_targets[target]; |
| lstate = NULL; |
| if (tstate != NULL) |
| lstate = tstate->enabled_luns[lun]; |
| |
| /* |
| * Commands for disabled luns go to the black hole driver. |
| */ |
| if (lstate == NULL) |
| lstate = ahd->black_hole; |
| |
| atio = (struct ccb_accept_tio*)SLIST_FIRST(&lstate->accept_tios); |
| if (atio == NULL) { |
| ahd->flags |= AHD_TQINFIFO_BLOCKED; |
| /* |
| * Wait for more ATIOs from the peripheral driver for this lun. |
| */ |
| return (1); |
| } else |
| ahd->flags &= ~AHD_TQINFIFO_BLOCKED; |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_TQIN) != 0) |
| printk("Incoming command from %d for %d:%d%s\n", |
| initiator, target, lun, |
| lstate == ahd->black_hole ? "(Black Holed)" : ""); |
| #endif |
| SLIST_REMOVE_HEAD(&lstate->accept_tios, sim_links.sle); |
| |
| if (lstate == ahd->black_hole) { |
| /* Fill in the wildcards */ |
| atio->ccb_h.target_id = target; |
| atio->ccb_h.target_lun = lun; |
| } |
| |
| /* |
| * Package it up and send it off to |
| * whomever has this lun enabled. |
| */ |
| atio->sense_len = 0; |
| atio->init_id = initiator; |
| if (byte[0] != 0xFF) { |
| /* Tag was included */ |
| atio->tag_action = *byte++; |
| atio->tag_id = *byte++; |
| atio->ccb_h.flags = CAM_TAG_ACTION_VALID; |
| } else { |
| atio->ccb_h.flags = 0; |
| } |
| byte++; |
| |
| /* Okay. Now determine the cdb size based on the command code */ |
| switch (*byte >> CMD_GROUP_CODE_SHIFT) { |
| case 0: |
| atio->cdb_len = 6; |
| break; |
| case 1: |
| case 2: |
| atio->cdb_len = 10; |
| break; |
| case 4: |
| atio->cdb_len = 16; |
| break; |
| case 5: |
| atio->cdb_len = 12; |
| break; |
| case 3: |
| default: |
| /* Only copy the opcode. */ |
| atio->cdb_len = 1; |
| printk("Reserved or VU command code type encountered\n"); |
| break; |
| } |
| |
| memcpy(atio->cdb_io.cdb_bytes, byte, atio->cdb_len); |
| |
| atio->ccb_h.status |= CAM_CDB_RECVD; |
| |
| if ((cmd->identify & MSG_IDENTIFY_DISCFLAG) == 0) { |
| /* |
| * We weren't allowed to disconnect. |
| * We're hanging on the bus until a |
| * continue target I/O comes in response |
| * to this accept tio. |
| */ |
| #ifdef AHD_DEBUG |
| if ((ahd_debug & AHD_SHOW_TQIN) != 0) |
| printk("Received Immediate Command %d:%d:%d - %p\n", |
| initiator, target, lun, ahd->pending_device); |
| #endif |
| ahd->pending_device = lstate; |
| ahd_freeze_ccb((union ccb *)atio); |
| atio->ccb_h.flags |= CAM_DIS_DISCONNECT; |
| } |
| xpt_done((union ccb*)atio); |
| return (0); |
| } |
| |
| #endif |