// SPDX-License-Identifier: GPL-2.0-only /* * QLogic qlcnic NIC Driver * Copyright (c) 2009-2013 QLogic Corporation */ #include #include "qlcnic.h" #include "qlcnic_hdr.h" #include "qlcnic_83xx_hw.h" #include "qlcnic_hw.h" #define QLC_83XX_MINIDUMP_FLASH 0x520000 #define QLC_83XX_OCM_INDEX 3 #define QLC_83XX_PCI_INDEX 0 #define QLC_83XX_DMA_ENGINE_INDEX 8 static const u32 qlcnic_ms_read_data[] = { 0x410000A8, 0x410000AC, 0x410000B8, 0x410000BC }; #define QLCNIC_DUMP_WCRB BIT_0 #define QLCNIC_DUMP_RWCRB BIT_1 #define QLCNIC_DUMP_ANDCRB BIT_2 #define QLCNIC_DUMP_ORCRB BIT_3 #define QLCNIC_DUMP_POLLCRB BIT_4 #define QLCNIC_DUMP_RD_SAVE BIT_5 #define QLCNIC_DUMP_WRT_SAVED BIT_6 #define QLCNIC_DUMP_MOD_SAVE_ST BIT_7 #define QLCNIC_DUMP_SKIP BIT_7 #define QLCNIC_DUMP_MASK_MAX 0xff struct qlcnic_pex_dma_descriptor { u32 read_data_size; u32 dma_desc_cmd; u32 src_addr_low; u32 src_addr_high; u32 dma_bus_addr_low; u32 dma_bus_addr_high; u32 rsvd[6]; } __packed; struct qlcnic_common_entry_hdr { u32 type; u32 offset; u32 cap_size; #if defined(__LITTLE_ENDIAN) u8 mask; u8 rsvd[2]; u8 flags; #else u8 flags; u8 rsvd[2]; u8 mask; #endif } __packed; struct __crb { u32 addr; #if defined(__LITTLE_ENDIAN) u8 stride; u8 rsvd1[3]; #else u8 rsvd1[3]; u8 stride; #endif u32 data_size; u32 no_ops; u32 rsvd2[4]; } __packed; struct __ctrl { u32 addr; #if defined(__LITTLE_ENDIAN) u8 stride; u8 index_a; u16 timeout; #else u16 timeout; u8 index_a; u8 stride; #endif u32 data_size; u32 no_ops; #if defined(__LITTLE_ENDIAN) u8 opcode; u8 index_v; u8 shl_val; u8 shr_val; #else u8 shr_val; u8 shl_val; u8 index_v; u8 opcode; #endif u32 val1; u32 val2; u32 val3; } __packed; struct __cache { u32 addr; #if defined(__LITTLE_ENDIAN) u16 stride; u16 init_tag_val; #else u16 init_tag_val; u16 stride; #endif u32 size; u32 no_ops; u32 ctrl_addr; u32 ctrl_val; u32 read_addr; #if defined(__LITTLE_ENDIAN) u8 read_addr_stride; u8 read_addr_num; u8 rsvd1[2]; #else u8 rsvd1[2]; u8 read_addr_num; u8 read_addr_stride; #endif } __packed; struct __ocm { u8 rsvd[8]; u32 size; u32 no_ops; u8 rsvd1[8]; u32 read_addr; u32 read_addr_stride; } __packed; struct __mem { u32 desc_card_addr; u32 dma_desc_cmd; u32 start_dma_cmd; u32 rsvd[3]; u32 addr; u32 size; } __packed; struct __mux { u32 addr; u8 rsvd[4]; u32 size; u32 no_ops; u32 val; u32 val_stride; u32 read_addr; u8 rsvd2[4]; } __packed; struct __queue { u32 sel_addr; #if defined(__LITTLE_ENDIAN) u16 stride; u8 rsvd[2]; #else u8 rsvd[2]; u16 stride; #endif u32 size; u32 no_ops; u8 rsvd2[8]; u32 read_addr; #if defined(__LITTLE_ENDIAN) u8 read_addr_stride; u8 read_addr_cnt; u8 rsvd3[2]; #else u8 rsvd3[2]; u8 read_addr_cnt; u8 read_addr_stride; #endif } __packed; struct __pollrd { u32 sel_addr; u32 read_addr; u32 sel_val; #if defined(__LITTLE_ENDIAN) u16 sel_val_stride; u16 no_ops; #else u16 no_ops; u16 sel_val_stride; #endif u32 poll_wait; u32 poll_mask; u32 data_size; u8 rsvd[4]; } __packed; struct __mux2 { u32 sel_addr1; u32 sel_addr2; u32 sel_val1; u32 sel_val2; u32 no_ops; u32 sel_val_mask; u32 read_addr; #if defined(__LITTLE_ENDIAN) u8 sel_val_stride; u8 data_size; u8 rsvd[2]; #else u8 rsvd[2]; u8 data_size; u8 sel_val_stride; #endif } __packed; struct __pollrdmwr { u32 addr1; u32 addr2; u32 val1; u32 val2; u32 poll_wait; u32 poll_mask; u32 mod_mask; u32 data_size; } __packed; struct qlcnic_dump_entry { struct qlcnic_common_entry_hdr hdr; union { struct __crb crb; struct __cache cache; struct __ocm ocm; struct __mem mem; struct __mux mux; struct __queue que; struct __ctrl ctrl; struct __pollrdmwr pollrdmwr; struct __mux2 mux2; struct __pollrd pollrd; } region; } __packed; enum qlcnic_minidump_opcode { QLCNIC_DUMP_NOP = 0, QLCNIC_DUMP_READ_CRB = 1, QLCNIC_DUMP_READ_MUX = 2, QLCNIC_DUMP_QUEUE = 3, QLCNIC_DUMP_BRD_CONFIG = 4, QLCNIC_DUMP_READ_OCM = 6, QLCNIC_DUMP_PEG_REG = 7, QLCNIC_DUMP_L1_DTAG = 8, QLCNIC_DUMP_L1_ITAG = 9, QLCNIC_DUMP_L1_DATA = 11, QLCNIC_DUMP_L1_INST = 12, QLCNIC_DUMP_L2_DTAG = 21, QLCNIC_DUMP_L2_ITAG = 22, QLCNIC_DUMP_L2_DATA = 23, QLCNIC_DUMP_L2_INST = 24, QLCNIC_DUMP_POLL_RD = 35, QLCNIC_READ_MUX2 = 36, QLCNIC_READ_POLLRDMWR = 37, QLCNIC_DUMP_READ_ROM = 71, QLCNIC_DUMP_READ_MEM = 72, QLCNIC_DUMP_READ_CTRL = 98, QLCNIC_DUMP_TLHDR = 99, QLCNIC_DUMP_RDEND = 255 }; inline u32 qlcnic_82xx_get_saved_state(void *t_hdr, u32 index) { struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr; return hdr->saved_state[index]; } inline void qlcnic_82xx_set_saved_state(void *t_hdr, u32 index, u32 value) { struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr; hdr->saved_state[index] = value; } void qlcnic_82xx_cache_tmpl_hdr_values(struct qlcnic_fw_dump *fw_dump) { struct qlcnic_82xx_dump_template_hdr *hdr; hdr = fw_dump->tmpl_hdr; fw_dump->tmpl_hdr_size = hdr->size; fw_dump->version = hdr->version; fw_dump->num_entries = hdr->num_entries; fw_dump->offset = hdr->offset; hdr->drv_cap_mask = hdr->cap_mask; fw_dump->cap_mask = hdr->cap_mask; fw_dump->use_pex_dma = (hdr->capabilities & BIT_0) ? true : false; } inline u32 qlcnic_82xx_get_cap_size(void *t_hdr, int index) { struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr; return hdr->cap_sizes[index]; } void qlcnic_82xx_set_sys_info(void *t_hdr, int idx, u32 value) { struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr; hdr->sys_info[idx] = value; } void qlcnic_82xx_store_cap_mask(void *tmpl_hdr, u32 mask) { struct qlcnic_82xx_dump_template_hdr *hdr = tmpl_hdr; hdr->drv_cap_mask = mask; } inline u32 qlcnic_83xx_get_saved_state(void *t_hdr, u32 index) { struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr; return hdr->saved_state[index]; } inline void qlcnic_83xx_set_saved_state(void *t_hdr, u32 index, u32 value) { struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr; hdr->saved_state[index] = value; } #define QLCNIC_TEMPLATE_VERSION (0x20001) void qlcnic_83xx_cache_tmpl_hdr_values(struct qlcnic_fw_dump *fw_dump) { struct qlcnic_83xx_dump_template_hdr *hdr; hdr = fw_dump->tmpl_hdr; fw_dump->tmpl_hdr_size = hdr->size; fw_dump->version = hdr->version; fw_dump->num_entries = hdr->num_entries; fw_dump->offset = hdr->offset; hdr->drv_cap_mask = hdr->cap_mask; fw_dump->cap_mask = hdr->cap_mask; fw_dump->use_pex_dma = (fw_dump->version & 0xfffff) >= QLCNIC_TEMPLATE_VERSION; } inline u32 qlcnic_83xx_get_cap_size(void *t_hdr, int index) { struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr; return hdr->cap_sizes[index]; } void qlcnic_83xx_set_sys_info(void *t_hdr, int idx, u32 value) { struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr; hdr->sys_info[idx] = value; } void qlcnic_83xx_store_cap_mask(void *tmpl_hdr, u32 mask) { struct qlcnic_83xx_dump_template_hdr *hdr; hdr = tmpl_hdr; hdr->drv_cap_mask = mask; } struct qlcnic_dump_operations { enum qlcnic_minidump_opcode opcode; u32 (*handler)(struct qlcnic_adapter *, struct qlcnic_dump_entry *, __le32 *); }; static u32 qlcnic_dump_crb(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { int i; u32 addr, data; struct __crb *crb = &entry->region.crb; addr = crb->addr; for (i = 0; i < crb->no_ops; i++) { data = qlcnic_ind_rd(adapter, addr); *buffer++ = cpu_to_le32(addr); *buffer++ = cpu_to_le32(data); addr += crb->stride; } return crb->no_ops * 2 * sizeof(u32); } static u32 qlcnic_dump_ctrl(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { void *hdr = adapter->ahw->fw_dump.tmpl_hdr; struct __ctrl *ctr = &entry->region.ctrl; int i, k, timeout = 0; u32 addr, data, temp; u8 no_ops; addr = ctr->addr; no_ops = ctr->no_ops; for (i = 0; i < no_ops; i++) { k = 0; for (k = 0; k < 8; k++) { if (!(ctr->opcode & (1 << k))) continue; switch (1 << k) { case QLCNIC_DUMP_WCRB: qlcnic_ind_wr(adapter, addr, ctr->val1); break; case QLCNIC_DUMP_RWCRB: data = qlcnic_ind_rd(adapter, addr); qlcnic_ind_wr(adapter, addr, data); break; case QLCNIC_DUMP_ANDCRB: data = qlcnic_ind_rd(adapter, addr); qlcnic_ind_wr(adapter, addr, (data & ctr->val2)); break; case QLCNIC_DUMP_ORCRB: data = qlcnic_ind_rd(adapter, addr); qlcnic_ind_wr(adapter, addr, (data | ctr->val3)); break; case QLCNIC_DUMP_POLLCRB: while (timeout <= ctr->timeout) { data = qlcnic_ind_rd(adapter, addr); if ((data & ctr->val2) == ctr->val1) break; usleep_range(1000, 2000); timeout++; } if (timeout > ctr->timeout) { dev_info(&adapter->pdev->dev, "Timed out, aborting poll CRB\n"); return -EINVAL; } break; case QLCNIC_DUMP_RD_SAVE: temp = ctr->index_a; if (temp) addr = qlcnic_get_saved_state(adapter, hdr, temp); data = qlcnic_ind_rd(adapter, addr); qlcnic_set_saved_state(adapter, hdr, ctr->index_v, data); break; case QLCNIC_DUMP_WRT_SAVED: temp = ctr->index_v; if (temp) data = qlcnic_get_saved_state(adapter, hdr, temp); else data = ctr->val1; temp = ctr->index_a; if (temp) addr = qlcnic_get_saved_state(adapter, hdr, temp); qlcnic_ind_wr(adapter, addr, data); break; case QLCNIC_DUMP_MOD_SAVE_ST: data = qlcnic_get_saved_state(adapter, hdr, ctr->index_v); data <<= ctr->shl_val; data >>= ctr->shr_val; if (ctr->val2) data &= ctr->val2; data |= ctr->val3; data += ctr->val1; qlcnic_set_saved_state(adapter, hdr, ctr->index_v, data); break; default: dev_info(&adapter->pdev->dev, "Unknown opcode\n"); break; } } addr += ctr->stride; } return 0; } static u32 qlcnic_dump_mux(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { int loop; u32 val, data = 0; struct __mux *mux = &entry->region.mux; val = mux->val; for (loop = 0; loop < mux->no_ops; loop++) { qlcnic_ind_wr(adapter, mux->addr, val); data = qlcnic_ind_rd(adapter, mux->read_addr); *buffer++ = cpu_to_le32(val); *buffer++ = cpu_to_le32(data); val += mux->val_stride; } return 2 * mux->no_ops * sizeof(u32); } static u32 qlcnic_dump_que(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { int i, loop; u32 cnt, addr, data, que_id = 0; struct __queue *que = &entry->region.que; addr = que->read_addr; cnt = que->read_addr_cnt; for (loop = 0; loop < que->no_ops; loop++) { qlcnic_ind_wr(adapter, que->sel_addr, que_id); addr = que->read_addr; for (i = 0; i < cnt; i++) { data = qlcnic_ind_rd(adapter, addr); *buffer++ = cpu_to_le32(data); addr += que->read_addr_stride; } que_id += que->stride; } return que->no_ops * cnt * sizeof(u32); } static u32 qlcnic_dump_ocm(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { int i; u32 data; void __iomem *addr; struct __ocm *ocm = &entry->region.ocm; addr = adapter->ahw->pci_base0 + ocm->read_addr; for (i = 0; i < ocm->no_ops; i++) { data = readl(addr); *buffer++ = cpu_to_le32(data); addr += ocm->read_addr_stride; } return ocm->no_ops * sizeof(u32); } static u32 qlcnic_read_rom(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { int i, count = 0; u32 fl_addr, size, val, lck_val, addr; struct __mem *rom = &entry->region.mem; fl_addr = rom->addr; size = rom->size / 4; lock_try: lck_val = QLC_SHARED_REG_RD32(adapter, QLCNIC_FLASH_LOCK); if (!lck_val && count < MAX_CTL_CHECK) { usleep_range(10000, 11000); count++; goto lock_try; } QLC_SHARED_REG_WR32(adapter, QLCNIC_FLASH_LOCK_OWNER, adapter->ahw->pci_func); for (i = 0; i < size; i++) { addr = fl_addr & 0xFFFF0000; qlcnic_ind_wr(adapter, FLASH_ROM_WINDOW, addr); addr = LSW(fl_addr) + FLASH_ROM_DATA; val = qlcnic_ind_rd(adapter, addr); fl_addr += 4; *buffer++ = cpu_to_le32(val); } QLC_SHARED_REG_RD32(adapter, QLCNIC_FLASH_UNLOCK); return rom->size; } static u32 qlcnic_dump_l1_cache(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { int i; u32 cnt, val, data, addr; struct __cache *l1 = &entry->region.cache; val = l1->init_tag_val; for (i = 0; i < l1->no_ops; i++) { qlcnic_ind_wr(adapter, l1->addr, val); qlcnic_ind_wr(adapter, l1->ctrl_addr, LSW(l1->ctrl_val)); addr = l1->read_addr; cnt = l1->read_addr_num; while (cnt) { data = qlcnic_ind_rd(adapter, addr); *buffer++ = cpu_to_le32(data); addr += l1->read_addr_stride; cnt--; } val += l1->stride; } return l1->no_ops * l1->read_addr_num * sizeof(u32); } static u32 qlcnic_dump_l2_cache(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { int i; u32 cnt, val, data, addr; u8 poll_mask, poll_to, time_out = 0; struct __cache *l2 = &entry->region.cache; val = l2->init_tag_val; poll_mask = LSB(MSW(l2->ctrl_val)); poll_to = MSB(MSW(l2->ctrl_val)); for (i = 0; i < l2->no_ops; i++) { qlcnic_ind_wr(adapter, l2->addr, val); if (LSW(l2->ctrl_val)) qlcnic_ind_wr(adapter, l2->ctrl_addr, LSW(l2->ctrl_val)); if (!poll_mask) goto skip_poll; do { data = qlcnic_ind_rd(adapter, l2->ctrl_addr); if (!(data & poll_mask)) break; usleep_range(1000, 2000); time_out++; } while (time_out <= poll_to); if (time_out > poll_to) { dev_err(&adapter->pdev->dev, "Timeout exceeded in %s, aborting dump\n", __func__); return -EINVAL; } skip_poll: addr = l2->read_addr; cnt = l2->read_addr_num; while (cnt) { data = qlcnic_ind_rd(adapter, addr); *buffer++ = cpu_to_le32(data); addr += l2->read_addr_stride; cnt--; } val += l2->stride; } return l2->no_ops * l2->read_addr_num * sizeof(u32); } static u32 qlcnic_read_memory_test_agent(struct qlcnic_adapter *adapter, struct __mem *mem, __le32 *buffer, int *ret) { u32 addr, data, test; int i, reg_read; reg_read = mem->size; addr = mem->addr; /* check for data size of multiple of 16 and 16 byte alignment */ if ((addr & 0xf) || (reg_read%16)) { dev_info(&adapter->pdev->dev, "Unaligned memory addr:0x%x size:0x%x\n", addr, reg_read); *ret = -EINVAL; return 0; } mutex_lock(&adapter->ahw->mem_lock); while (reg_read != 0) { qlcnic_ind_wr(adapter, QLCNIC_MS_ADDR_LO, addr); qlcnic_ind_wr(adapter, QLCNIC_MS_ADDR_HI, 0); qlcnic_ind_wr(adapter, QLCNIC_MS_CTRL, QLCNIC_TA_START_ENABLE); for (i = 0; i < MAX_CTL_CHECK; i++) { test = qlcnic_ind_rd(adapter, QLCNIC_MS_CTRL); if (!(test & TA_CTL_BUSY)) break; } if (i == MAX_CTL_CHECK) { if (printk_ratelimit()) { dev_err(&adapter->pdev->dev, "failed to read through agent\n"); *ret = -EIO; goto out; } } for (i = 0; i < 4; i++) { data = qlcnic_ind_rd(adapter, qlcnic_ms_read_data[i]); *buffer++ = cpu_to_le32(data); } addr += 16; reg_read -= 16; ret += 16; cond_resched(); } out: mutex_unlock(&adapter->ahw->mem_lock); return mem->size; } /* DMA register base address */ #define QLC_DMA_REG_BASE_ADDR(dma_no) (0x77320000 + (dma_no * 0x10000)) /* DMA register offsets w.r.t base address */ #define QLC_DMA_CMD_BUFF_ADDR_LOW 0 #define QLC_DMA_CMD_BUFF_ADDR_HI 4 #define QLC_DMA_CMD_STATUS_CTRL 8 static int qlcnic_start_pex_dma(struct qlcnic_adapter *adapter, struct __mem *mem) { struct device *dev = &adapter->pdev->dev; u32 dma_no, dma_base_addr, temp_addr; int i, ret, dma_sts; void *tmpl_hdr; tmpl_hdr = adapter->ahw->fw_dump.tmpl_hdr; dma_no = qlcnic_get_saved_state(adapter, tmpl_hdr, QLC_83XX_DMA_ENGINE_INDEX); dma_base_addr = QLC_DMA_REG_BASE_ADDR(dma_no); temp_addr = dma_base_addr + QLC_DMA_CMD_BUFF_ADDR_LOW; ret = qlcnic_ind_wr(adapter, temp_addr, mem->desc_card_addr); if (ret) return ret; temp_addr = dma_base_addr + QLC_DMA_CMD_BUFF_ADDR_HI; ret = qlcnic_ind_wr(adapter, temp_addr, 0); if (ret) return ret; temp_addr = dma_base_addr + QLC_DMA_CMD_STATUS_CTRL; ret = qlcnic_ind_wr(adapter, temp_addr, mem->start_dma_cmd); if (ret) return ret; /* Wait for DMA to complete */ temp_addr = dma_base_addr + QLC_DMA_CMD_STATUS_CTRL; for (i = 0; i < 400; i++) { dma_sts = qlcnic_ind_rd(adapter, temp_addr); if (dma_sts & BIT_1) usleep_range(250, 500); else break; } if (i >= 400) { dev_info(dev, "PEX DMA operation timed out"); ret = -EIO; } return ret; } static u32 qlcnic_read_memory_pexdma(struct qlcnic_adapter *adapter, struct __mem *mem, __le32 *buffer, int *ret) { struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump; u32 temp, dma_base_addr, size = 0, read_size = 0; struct qlcnic_pex_dma_descriptor *dma_descr; struct device *dev = &adapter->pdev->dev; dma_addr_t dma_phys_addr; void *dma_buffer; void *tmpl_hdr; tmpl_hdr = fw_dump->tmpl_hdr; /* Check if DMA engine is available */ temp = qlcnic_get_saved_state(adapter, tmpl_hdr, QLC_83XX_DMA_ENGINE_INDEX); dma_base_addr = QLC_DMA_REG_BASE_ADDR(temp); temp = qlcnic_ind_rd(adapter, dma_base_addr + QLC_DMA_CMD_STATUS_CTRL); if (!(temp & BIT_31)) { dev_info(dev, "%s: DMA engine is not available\n", __func__); *ret = -EIO; return 0; } /* Create DMA descriptor */ dma_descr = kzalloc(sizeof(struct qlcnic_pex_dma_descriptor), GFP_KERNEL); if (!dma_descr) { *ret = -ENOMEM; return 0; } /* dma_desc_cmd 0:15 = 0 * dma_desc_cmd 16:19 = mem->dma_desc_cmd 0:3 * dma_desc_cmd 20:23 = pci function number * dma_desc_cmd 24:31 = mem->dma_desc_cmd 8:15 */ dma_phys_addr = fw_dump->phys_addr; dma_buffer = fw_dump->dma_buffer; temp = 0; temp = mem->dma_desc_cmd & 0xff0f; temp |= (adapter->ahw->pci_func & 0xf) << 4; dma_descr->dma_desc_cmd = (temp << 16) & 0xffff0000; dma_descr->dma_bus_addr_low = LSD(dma_phys_addr); dma_descr->dma_bus_addr_high = MSD(dma_phys_addr); dma_descr->src_addr_high = 0; /* Collect memory dump using multiple DMA operations if required */ while (read_size < mem->size) { if (mem->size - read_size >= QLC_PEX_DMA_READ_SIZE) size = QLC_PEX_DMA_READ_SIZE; else size = mem->size - read_size; dma_descr->src_addr_low = mem->addr + read_size; dma_descr->read_data_size = size; /* Write DMA descriptor to MS memory*/ temp = sizeof(struct qlcnic_pex_dma_descriptor) / 16; *ret = qlcnic_ms_mem_write128(adapter, mem->desc_card_addr, (u32 *)dma_descr, temp); if (*ret) { dev_info(dev, "Failed to write DMA descriptor to MS memory at address 0x%x\n", mem->desc_card_addr); goto free_dma_descr; } *ret = qlcnic_start_pex_dma(adapter, mem); if (*ret) { dev_info(dev, "Failed to start PEX DMA operation\n"); goto free_dma_descr; } memcpy(buffer, dma_buffer, size); buffer += size / 4; read_size += size; } free_dma_descr: kfree(dma_descr); return read_size; } static u32 qlcnic_read_memory(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump; struct device *dev = &adapter->pdev->dev; struct __mem *mem = &entry->region.mem; u32 data_size; int ret = 0; if (fw_dump->use_pex_dma) { data_size = qlcnic_read_memory_pexdma(adapter, mem, buffer, &ret); if (ret) dev_info(dev, "Failed to read memory dump using PEX DMA: mask[0x%x]\n", entry->hdr.mask); else return data_size; } data_size = qlcnic_read_memory_test_agent(adapter, mem, buffer, &ret); if (ret) { dev_info(dev, "Failed to read memory dump using test agent method: mask[0x%x]\n", entry->hdr.mask); return 0; } else { return data_size; } } static u32 qlcnic_dump_nop(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { entry->hdr.flags |= QLCNIC_DUMP_SKIP; return 0; } static int qlcnic_valid_dump_entry(struct device *dev, struct qlcnic_dump_entry *entry, u32 size) { int ret = 1; if (size != entry->hdr.cap_size) { dev_err(dev, "Invalid entry, Type:%d\tMask:%d\tSize:%dCap_size:%d\n", entry->hdr.type, entry->hdr.mask, size, entry->hdr.cap_size); ret = 0; } return ret; } static u32 qlcnic_read_pollrdmwr(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { struct __pollrdmwr *poll = &entry->region.pollrdmwr; u32 data, wait_count, poll_wait, temp; poll_wait = poll->poll_wait; qlcnic_ind_wr(adapter, poll->addr1, poll->val1); wait_count = 0; while (wait_count < poll_wait) { data = qlcnic_ind_rd(adapter, poll->addr1); if ((data & poll->poll_mask) != 0) break; wait_count++; } if (wait_count == poll_wait) { dev_err(&adapter->pdev->dev, "Timeout exceeded in %s, aborting dump\n", __func__); return 0; } data = qlcnic_ind_rd(adapter, poll->addr2) & poll->mod_mask; qlcnic_ind_wr(adapter, poll->addr2, data); qlcnic_ind_wr(adapter, poll->addr1, poll->val2); wait_count = 0; while (wait_count < poll_wait) { temp = qlcnic_ind_rd(adapter, poll->addr1); if ((temp & poll->poll_mask) != 0) break; wait_count++; } *buffer++ = cpu_to_le32(poll->addr2); *buffer++ = cpu_to_le32(data); return 2 * sizeof(u32); } static u32 qlcnic_read_pollrd(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { struct __pollrd *pollrd = &entry->region.pollrd; u32 data, wait_count, poll_wait, sel_val; int i; poll_wait = pollrd->poll_wait; sel_val = pollrd->sel_val; for (i = 0; i < pollrd->no_ops; i++) { qlcnic_ind_wr(adapter, pollrd->sel_addr, sel_val); wait_count = 0; while (wait_count < poll_wait) { data = qlcnic_ind_rd(adapter, pollrd->sel_addr); if ((data & pollrd->poll_mask) != 0) break; wait_count++; } if (wait_count == poll_wait) { dev_err(&adapter->pdev->dev, "Timeout exceeded in %s, aborting dump\n", __func__); return 0; } data = qlcnic_ind_rd(adapter, pollrd->read_addr); *buffer++ = cpu_to_le32(sel_val); *buffer++ = cpu_to_le32(data); sel_val += pollrd->sel_val_stride; } return pollrd->no_ops * (2 * sizeof(u32)); } static u32 qlcnic_read_mux2(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { struct __mux2 *mux2 = &entry->region.mux2; u32 data; u32 t_sel_val, sel_val1, sel_val2; int i; sel_val1 = mux2->sel_val1; sel_val2 = mux2->sel_val2; for (i = 0; i < mux2->no_ops; i++) { qlcnic_ind_wr(adapter, mux2->sel_addr1, sel_val1); t_sel_val = sel_val1 & mux2->sel_val_mask; qlcnic_ind_wr(adapter, mux2->sel_addr2, t_sel_val); data = qlcnic_ind_rd(adapter, mux2->read_addr); *buffer++ = cpu_to_le32(t_sel_val); *buffer++ = cpu_to_le32(data); qlcnic_ind_wr(adapter, mux2->sel_addr1, sel_val2); t_sel_val = sel_val2 & mux2->sel_val_mask; qlcnic_ind_wr(adapter, mux2->sel_addr2, t_sel_val); data = qlcnic_ind_rd(adapter, mux2->read_addr); *buffer++ = cpu_to_le32(t_sel_val); *buffer++ = cpu_to_le32(data); sel_val1 += mux2->sel_val_stride; sel_val2 += mux2->sel_val_stride; } return mux2->no_ops * (4 * sizeof(u32)); } static u32 qlcnic_83xx_dump_rom(struct qlcnic_adapter *adapter, struct qlcnic_dump_entry *entry, __le32 *buffer) { u32 fl_addr, size; struct __mem *rom = &entry->region.mem; fl_addr = rom->addr; size = rom->size / 4; if (!qlcnic_83xx_lockless_flash_read32(adapter, fl_addr, (u8 *)buffer, size)) return rom->size; return 0; } static const struct qlcnic_dump_operations qlcnic_fw_dump_ops[] = { {QLCNIC_DUMP_NOP, qlcnic_dump_nop}, {QLCNIC_DUMP_READ_CRB, qlcnic_dump_crb}, {QLCNIC_DUMP_READ_MUX, qlcnic_dump_mux}, {QLCNIC_DUMP_QUEUE, qlcnic_dump_que}, {QLCNIC_DUMP_BRD_CONFIG, qlcnic_read_rom}, {QLCNIC_DUMP_READ_OCM, qlcnic_dump_ocm}, {QLCNIC_DUMP_PEG_REG, qlcnic_dump_ctrl}, {QLCNIC_DUMP_L1_DTAG, qlcnic_dump_l1_cache}, {QLCNIC_DUMP_L1_ITAG, qlcnic_dump_l1_cache}, {QLCNIC_DUMP_L1_DATA, qlcnic_dump_l1_cache}, {QLCNIC_DUMP_L1_INST, qlcnic_dump_l1_cache}, {QLCNIC_DUMP_L2_DTAG, qlcnic_dump_l2_cache}, {QLCNIC_DUMP_L2_ITAG, qlcnic_dump_l2_cache}, {QLCNIC_DUMP_L2_DATA, qlcnic_dump_l2_cache}, {QLCNIC_DUMP_L2_INST, qlcnic_dump_l2_cache}, {QLCNIC_DUMP_READ_ROM, qlcnic_read_rom}, {QLCNIC_DUMP_READ_MEM, qlcnic_read_memory}, {QLCNIC_DUMP_READ_CTRL, qlcnic_dump_ctrl}, {QLCNIC_DUMP_TLHDR, qlcnic_dump_nop}, {QLCNIC_DUMP_RDEND, qlcnic_dump_nop}, }; static const struct qlcnic_dump_operations qlcnic_83xx_fw_dump_ops[] = { {QLCNIC_DUMP_NOP, qlcnic_dump_nop}, {QLCNIC_DUMP_READ_CRB, qlcnic_dump_crb}, {QLCNIC_DUMP_READ_MUX, qlcnic_dump_mux}, {QLCNIC_DUMP_QUEUE, qlcnic_dump_que}, {QLCNIC_DUMP_BRD_CONFIG, qlcnic_83xx_dump_rom}, {QLCNIC_DUMP_READ_OCM, qlcnic_dump_ocm}, {QLCNIC_DUMP_PEG_REG, qlcnic_dump_ctrl}, {QLCNIC_DUMP_L1_DTAG, qlcnic_dump_l1_cache}, {QLCNIC_DUMP_L1_ITAG, qlcnic_dump_l1_cache}, {QLCNIC_DUMP_L1_DATA, qlcnic_dump_l1_cache}, {QLCNIC_DUMP_L1_INST, qlcnic_dump_l1_cache}, {QLCNIC_DUMP_L2_DTAG, qlcnic_dump_l2_cache}, {QLCNIC_DUMP_L2_ITAG, qlcnic_dump_l2_cache}, {QLCNIC_DUMP_L2_DATA, qlcnic_dump_l2_cache}, {QLCNIC_DUMP_L2_INST, qlcnic_dump_l2_cache}, {QLCNIC_DUMP_POLL_RD, qlcnic_read_pollrd}, {QLCNIC_READ_MUX2, qlcnic_read_mux2}, {QLCNIC_READ_POLLRDMWR, qlcnic_read_pollrdmwr}, {QLCNIC_DUMP_READ_ROM, qlcnic_83xx_dump_rom}, {QLCNIC_DUMP_READ_MEM, qlcnic_read_memory}, {QLCNIC_DUMP_READ_CTRL, qlcnic_dump_ctrl}, {QLCNIC_DUMP_TLHDR, qlcnic_dump_nop}, {QLCNIC_DUMP_RDEND, qlcnic_dump_nop}, }; static uint32_t qlcnic_temp_checksum(uint32_t *temp_buffer, u32 temp_size) { uint64_t sum = 0; int count = temp_size / sizeof(uint32_t); while (count-- > 0) sum += *temp_buffer++; while (sum >> 32) sum = (sum & 0xFFFFFFFF) + (sum >> 32); return ~sum; } static int qlcnic_fw_flash_get_minidump_temp(struct qlcnic_adapter *adapter, u8 *buffer, u32 size) { int ret = 0; if (qlcnic_82xx_check(adapter)) return -EIO; if (qlcnic_83xx_lock_flash(adapter)) return -EIO; ret = qlcnic_83xx_lockless_flash_read32(adapter, QLC_83XX_MINIDUMP_FLASH, buffer, size / sizeof(u32)); qlcnic_83xx_unlock_flash(adapter); return ret; } static int qlcnic_fw_flash_get_minidump_temp_size(struct qlcnic_adapter *adapter, struct qlcnic_cmd_args *cmd) { struct qlcnic_83xx_dump_template_hdr tmp_hdr; u32 size = sizeof(tmp_hdr) / sizeof(u32); int ret = 0; if (qlcnic_82xx_check(adapter)) return -EIO; if (qlcnic_83xx_lock_flash(adapter)) return -EIO; ret = qlcnic_83xx_lockless_flash_read32(adapter, QLC_83XX_MINIDUMP_FLASH, (u8 *)&tmp_hdr, size); qlcnic_83xx_unlock_flash(adapter); cmd->rsp.arg[2] = tmp_hdr.size; cmd->rsp.arg[3] = tmp_hdr.version; return ret; } static int qlcnic_fw_get_minidump_temp_size(struct qlcnic_adapter *adapter, u32 *version, u32 *temp_size, u8 *use_flash_temp) { int err = 0; struct qlcnic_cmd_args cmd; if (qlcnic_alloc_mbx_args(&cmd, adapter, QLCNIC_CMD_TEMP_SIZE)) return -ENOMEM; err = qlcnic_issue_cmd(adapter, &cmd); if (err != QLCNIC_RCODE_SUCCESS) { if (qlcnic_fw_flash_get_minidump_temp_size(adapter, &cmd)) { qlcnic_free_mbx_args(&cmd); return -EIO; } *use_flash_temp = 1; } *temp_size = cmd.rsp.arg[2]; *version = cmd.rsp.arg[3]; qlcnic_free_mbx_args(&cmd); if (!(*temp_size)) return -EIO; return 0; } static int __qlcnic_fw_cmd_get_minidump_temp(struct qlcnic_adapter *adapter, u32 *buffer, u32 temp_size) { int err = 0, i; void *tmp_addr; __le32 *tmp_buf; struct qlcnic_cmd_args cmd; dma_addr_t tmp_addr_t = 0; tmp_addr = dma_alloc_coherent(&adapter->pdev->dev, temp_size, &tmp_addr_t, GFP_KERNEL); if (!tmp_addr) return -ENOMEM; if (qlcnic_alloc_mbx_args(&cmd, adapter, QLCNIC_CMD_GET_TEMP_HDR)) { err = -ENOMEM; goto free_mem; } cmd.req.arg[1] = LSD(tmp_addr_t); cmd.req.arg[2] = MSD(tmp_addr_t); cmd.req.arg[3] = temp_size; err = qlcnic_issue_cmd(adapter, &cmd); tmp_buf = tmp_addr; if (err == QLCNIC_RCODE_SUCCESS) { for (i = 0; i < temp_size / sizeof(u32); i++) *buffer++ = __le32_to_cpu(*tmp_buf++); } qlcnic_free_mbx_args(&cmd); free_mem: dma_free_coherent(&adapter->pdev->dev, temp_size, tmp_addr, tmp_addr_t); return err; } int qlcnic_fw_cmd_get_minidump_temp(struct qlcnic_adapter *adapter) { struct qlcnic_hardware_context *ahw; struct qlcnic_fw_dump *fw_dump; u32 version, csum, *tmp_buf; u8 use_flash_temp = 0; u32 temp_size = 0; void *temp_buffer; int err; ahw = adapter->ahw; fw_dump = &ahw->fw_dump; err = qlcnic_fw_get_minidump_temp_size(adapter, &version, &temp_size, &use_flash_temp); if (err) { dev_err(&adapter->pdev->dev, "Can't get template size %d\n", err); return -EIO; } fw_dump->tmpl_hdr = vzalloc(temp_size); if (!fw_dump->tmpl_hdr) return -ENOMEM; tmp_buf = (u32 *)fw_dump->tmpl_hdr; if (use_flash_temp) goto flash_temp; err = __qlcnic_fw_cmd_get_minidump_temp(adapter, tmp_buf, temp_size); if (err) { flash_temp: err = qlcnic_fw_flash_get_minidump_temp(adapter, (u8 *)tmp_buf, temp_size); if (err) { dev_err(&adapter->pdev->dev, "Failed to get minidump template header %d\n", err); vfree(fw_dump->tmpl_hdr); fw_dump->tmpl_hdr = NULL; return -EIO; } } csum = qlcnic_temp_checksum((uint32_t *)tmp_buf, temp_size); if (csum) { dev_err(&adapter->pdev->dev, "Template header checksum validation failed\n"); vfree(fw_dump->tmpl_hdr); fw_dump->tmpl_hdr = NULL; return -EIO; } qlcnic_cache_tmpl_hdr_values(adapter, fw_dump); if (fw_dump->use_pex_dma) { fw_dump->dma_buffer = NULL; temp_buffer = dma_alloc_coherent(&adapter->pdev->dev, QLC_PEX_DMA_READ_SIZE, &fw_dump->phys_addr, GFP_KERNEL); if (!temp_buffer) fw_dump->use_pex_dma = false; else fw_dump->dma_buffer = temp_buffer; } dev_info(&adapter->pdev->dev, "Default minidump capture mask 0x%x\n", fw_dump->cap_mask); qlcnic_enable_fw_dump_state(adapter); return 0; } int qlcnic_dump_fw(struct qlcnic_adapter *adapter) { struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump; const struct qlcnic_dump_operations *fw_dump_ops; struct qlcnic_83xx_dump_template_hdr *hdr_83xx; u32 entry_offset, dump, no_entries, buf_offset = 0; int i, k, ops_cnt, ops_index, dump_size = 0; struct device *dev = &adapter->pdev->dev; struct qlcnic_hardware_context *ahw; struct qlcnic_dump_entry *entry; void *tmpl_hdr; u32 ocm_window; __le32 *buffer; char mesg[64]; char *msg[] = {mesg, NULL}; ahw = adapter->ahw; tmpl_hdr = fw_dump->tmpl_hdr; /* Return if we don't have firmware dump template header */ if (!tmpl_hdr) return -EIO; if (!qlcnic_check_fw_dump_state(adapter)) { dev_info(&adapter->pdev->dev, "Dump not enabled\n"); return -EIO; } if (fw_dump->clr) { dev_info(&adapter->pdev->dev, "Previous dump not cleared, not capturing dump\n"); return -EIO; } netif_info(adapter->ahw, drv, adapter->netdev, "Take FW dump\n"); /* Calculate the size for dump data area only */ for (i = 2, k = 1; (i & QLCNIC_DUMP_MASK_MAX); i <<= 1, k++) if (i & fw_dump->cap_mask) dump_size += qlcnic_get_cap_size(adapter, tmpl_hdr, k); if (!dump_size) return -EIO; fw_dump->data = vzalloc(dump_size); if (!fw_dump->data) return -ENOMEM; buffer = fw_dump->data; fw_dump->size = dump_size; no_entries = fw_dump->num_entries; entry_offset = fw_dump->offset; qlcnic_set_sys_info(adapter, tmpl_hdr, 0, QLCNIC_DRIVER_VERSION); qlcnic_set_sys_info(adapter, tmpl_hdr, 1, adapter->fw_version); if (qlcnic_82xx_check(adapter)) { ops_cnt = ARRAY_SIZE(qlcnic_fw_dump_ops); fw_dump_ops = qlcnic_fw_dump_ops; } else { hdr_83xx = tmpl_hdr; ops_cnt = ARRAY_SIZE(qlcnic_83xx_fw_dump_ops); fw_dump_ops = qlcnic_83xx_fw_dump_ops; ocm_window = hdr_83xx->ocm_wnd_reg[ahw->pci_func]; hdr_83xx->saved_state[QLC_83XX_OCM_INDEX] = ocm_window; hdr_83xx->saved_state[QLC_83XX_PCI_INDEX] = ahw->pci_func; } for (i = 0; i < no_entries; i++) { entry = tmpl_hdr + entry_offset; if (!(entry->hdr.mask & fw_dump->cap_mask)) { entry->hdr.flags |= QLCNIC_DUMP_SKIP; entry_offset += entry->hdr.offset; continue; } /* Find the handler for this entry */ ops_index = 0; while (ops_index < ops_cnt) { if (entry->hdr.type == fw_dump_ops[ops_index].opcode) break; ops_index++; } if (ops_index == ops_cnt) { dev_info(dev, "Skipping unknown entry opcode %d\n", entry->hdr.type); entry->hdr.flags |= QLCNIC_DUMP_SKIP; entry_offset += entry->hdr.offset; continue; } /* Collect dump for this entry */ dump = fw_dump_ops[ops_index].handler(adapter, entry, buffer); if (!qlcnic_valid_dump_entry(dev, entry, dump)) { entry->hdr.flags |= QLCNIC_DUMP_SKIP; entry_offset += entry->hdr.offset; continue; } buf_offset += entry->hdr.cap_size; entry_offset += entry->hdr.offset; buffer = fw_dump->data + buf_offset; cond_resched(); } fw_dump->clr = 1; snprintf(mesg, sizeof(mesg), "FW_DUMP=%s", adapter->netdev->name); netdev_info(adapter->netdev, "Dump data %d bytes captured, dump data address = %p, template header size %d bytes, template address = %p\n", fw_dump->size, fw_dump->data, fw_dump->tmpl_hdr_size, fw_dump->tmpl_hdr); /* Send a udev event to notify availability of FW dump */ kobject_uevent_env(&dev->kobj, KOBJ_CHANGE, msg); return 0; } static inline bool qlcnic_83xx_md_check_extended_dump_capability(struct qlcnic_adapter *adapter) { /* For special adapters (with 0x8830 device ID), where iSCSI firmware * dump needs to be captured as part of regular firmware dump * collection process, firmware exports it's capability through * capability registers */ return ((adapter->pdev->device == PCI_DEVICE_ID_QLOGIC_QLE8830) && (adapter->ahw->extra_capability[0] & QLCNIC_FW_CAPABILITY_2_EXT_ISCSI_DUMP)); } void qlcnic_83xx_get_minidump_template(struct qlcnic_adapter *adapter) { u32 prev_version, current_version; struct qlcnic_hardware_context *ahw = adapter->ahw; struct qlcnic_fw_dump *fw_dump = &ahw->fw_dump; struct pci_dev *pdev = adapter->pdev; bool extended = false; int ret; prev_version = adapter->fw_version; current_version = qlcnic_83xx_get_fw_version(adapter); if (fw_dump->tmpl_hdr == NULL || current_version > prev_version) { vfree(fw_dump->tmpl_hdr); if (qlcnic_83xx_md_check_extended_dump_capability(adapter)) extended = !qlcnic_83xx_extend_md_capab(adapter); ret = qlcnic_fw_cmd_get_minidump_temp(adapter); if (ret) return; dev_info(&pdev->dev, "Supports FW dump capability\n"); /* Once we have minidump template with extended iSCSI dump * capability, update the minidump capture mask to 0x1f as * per FW requirement */ if (extended) { struct qlcnic_83xx_dump_template_hdr *hdr; hdr = fw_dump->tmpl_hdr; hdr->drv_cap_mask = 0x1f; fw_dump->cap_mask = 0x1f; dev_info(&pdev->dev, "Extended iSCSI dump capability and updated capture mask to 0x1f\n"); } } }