// SPDX-License-Identifier: GPL-2.0-only /* * Microsemi Switchtec(tm) PCIe Management Driver * Copyright (c) 2017, Microsemi Corporation */ #include #include #include #include #include #include #include #include MODULE_DESCRIPTION("Microsemi Switchtec(tm) NTB Driver"); MODULE_VERSION("0.1"); MODULE_LICENSE("GPL"); MODULE_AUTHOR("Microsemi Corporation"); static ulong max_mw_size = SZ_2M; module_param(max_mw_size, ulong, 0644); MODULE_PARM_DESC(max_mw_size, "Max memory window size reported to the upper layer"); static bool use_lut_mws; module_param(use_lut_mws, bool, 0644); MODULE_PARM_DESC(use_lut_mws, "Enable the use of the LUT based memory windows"); #define SWITCHTEC_NTB_MAGIC 0x45CC0001 #define MAX_MWS 128 struct shared_mw { u32 magic; u32 link_sta; u32 partition_id; u64 mw_sizes[MAX_MWS]; u32 spad[128]; }; #define MAX_DIRECT_MW ARRAY_SIZE(((struct ntb_ctrl_regs *)(0))->bar_entry) #define LUT_SIZE SZ_64K struct switchtec_ntb { struct ntb_dev ntb; struct switchtec_dev *stdev; int self_partition; int peer_partition; int doorbell_irq; int message_irq; struct ntb_info_regs __iomem *mmio_ntb; struct ntb_ctrl_regs __iomem *mmio_ctrl; struct ntb_dbmsg_regs __iomem *mmio_dbmsg; struct ntb_ctrl_regs __iomem *mmio_self_ctrl; struct ntb_ctrl_regs __iomem *mmio_peer_ctrl; struct ntb_dbmsg_regs __iomem *mmio_self_dbmsg; struct ntb_dbmsg_regs __iomem *mmio_peer_dbmsg; void __iomem *mmio_xlink_win; struct shared_mw *self_shared; struct shared_mw __iomem *peer_shared; dma_addr_t self_shared_dma; u64 db_mask; u64 db_valid_mask; int db_shift; int db_peer_shift; /* synchronize rmw access of db_mask and hw reg */ spinlock_t db_mask_lock; int nr_direct_mw; int nr_lut_mw; int nr_rsvd_luts; int direct_mw_to_bar[MAX_DIRECT_MW]; int peer_nr_direct_mw; int peer_nr_lut_mw; int peer_direct_mw_to_bar[MAX_DIRECT_MW]; bool link_is_up; enum ntb_speed link_speed; enum ntb_width link_width; struct work_struct check_link_status_work; bool link_force_down; }; static struct switchtec_ntb *ntb_sndev(struct ntb_dev *ntb) { return container_of(ntb, struct switchtec_ntb, ntb); } static int switchtec_ntb_part_op(struct switchtec_ntb *sndev, struct ntb_ctrl_regs __iomem *ctl, u32 op, int wait_status) { static const char * const op_text[] = { [NTB_CTRL_PART_OP_LOCK] = "lock", [NTB_CTRL_PART_OP_CFG] = "configure", [NTB_CTRL_PART_OP_RESET] = "reset", }; int i; u32 ps; int status; switch (op) { case NTB_CTRL_PART_OP_LOCK: status = NTB_CTRL_PART_STATUS_LOCKING; break; case NTB_CTRL_PART_OP_CFG: status = NTB_CTRL_PART_STATUS_CONFIGURING; break; case NTB_CTRL_PART_OP_RESET: status = NTB_CTRL_PART_STATUS_RESETTING; break; default: return -EINVAL; } iowrite32(op, &ctl->partition_op); for (i = 0; i < 1000; i++) { if (msleep_interruptible(50) != 0) { iowrite32(NTB_CTRL_PART_OP_RESET, &ctl->partition_op); return -EINTR; } ps = ioread32(&ctl->partition_status) & 0xFFFF; if (ps != status) break; } if (ps == wait_status) return 0; if (ps == status) { dev_err(&sndev->stdev->dev, "Timed out while performing %s (%d). (%08x)\n", op_text[op], op, ioread32(&ctl->partition_status)); return -ETIMEDOUT; } return -EIO; } static int switchtec_ntb_send_msg(struct switchtec_ntb *sndev, int idx, u32 val) { if (idx < 0 || idx >= ARRAY_SIZE(sndev->mmio_peer_dbmsg->omsg)) return -EINVAL; iowrite32(val, &sndev->mmio_peer_dbmsg->omsg[idx].msg); return 0; } static int switchtec_ntb_mw_count(struct ntb_dev *ntb, int pidx) { struct switchtec_ntb *sndev = ntb_sndev(ntb); int nr_direct_mw = sndev->peer_nr_direct_mw; int nr_lut_mw = sndev->peer_nr_lut_mw - sndev->nr_rsvd_luts; if (pidx != NTB_DEF_PEER_IDX) return -EINVAL; if (!use_lut_mws) nr_lut_mw = 0; return nr_direct_mw + nr_lut_mw; } static int lut_index(struct switchtec_ntb *sndev, int mw_idx) { return mw_idx - sndev->nr_direct_mw + sndev->nr_rsvd_luts; } static int peer_lut_index(struct switchtec_ntb *sndev, int mw_idx) { return mw_idx - sndev->peer_nr_direct_mw + sndev->nr_rsvd_luts; } static int switchtec_ntb_mw_get_align(struct ntb_dev *ntb, int pidx, int widx, resource_size_t *addr_align, resource_size_t *size_align, resource_size_t *size_max) { struct switchtec_ntb *sndev = ntb_sndev(ntb); int lut; resource_size_t size; if (pidx != NTB_DEF_PEER_IDX) return -EINVAL; lut = widx >= sndev->peer_nr_direct_mw; size = ioread64(&sndev->peer_shared->mw_sizes[widx]); if (size == 0) return -EINVAL; if (addr_align) *addr_align = lut ? size : SZ_4K; if (size_align) *size_align = lut ? size : SZ_4K; if (size_max) *size_max = size; return 0; } static void switchtec_ntb_mw_clr_direct(struct switchtec_ntb *sndev, int idx) { struct ntb_ctrl_regs __iomem *ctl = sndev->mmio_peer_ctrl; int bar = sndev->peer_direct_mw_to_bar[idx]; u32 ctl_val; ctl_val = ioread32(&ctl->bar_entry[bar].ctl); ctl_val &= ~NTB_CTRL_BAR_DIR_WIN_EN; iowrite32(ctl_val, &ctl->bar_entry[bar].ctl); iowrite32(0, &ctl->bar_entry[bar].win_size); iowrite32(0, &ctl->bar_ext_entry[bar].win_size); iowrite64(sndev->self_partition, &ctl->bar_entry[bar].xlate_addr); } static void switchtec_ntb_mw_clr_lut(struct switchtec_ntb *sndev, int idx) { struct ntb_ctrl_regs __iomem *ctl = sndev->mmio_peer_ctrl; iowrite64(0, &ctl->lut_entry[peer_lut_index(sndev, idx)]); } static void switchtec_ntb_mw_set_direct(struct switchtec_ntb *sndev, int idx, dma_addr_t addr, resource_size_t size) { int xlate_pos = ilog2(size); int bar = sndev->peer_direct_mw_to_bar[idx]; struct ntb_ctrl_regs __iomem *ctl = sndev->mmio_peer_ctrl; u32 ctl_val; ctl_val = ioread32(&ctl->bar_entry[bar].ctl); ctl_val |= NTB_CTRL_BAR_DIR_WIN_EN; iowrite32(ctl_val, &ctl->bar_entry[bar].ctl); iowrite32(xlate_pos | (lower_32_bits(size) & 0xFFFFF000), &ctl->bar_entry[bar].win_size); iowrite32(upper_32_bits(size), &ctl->bar_ext_entry[bar].win_size); iowrite64(sndev->self_partition | addr, &ctl->bar_entry[bar].xlate_addr); } static void switchtec_ntb_mw_set_lut(struct switchtec_ntb *sndev, int idx, dma_addr_t addr, resource_size_t size) { struct ntb_ctrl_regs __iomem *ctl = sndev->mmio_peer_ctrl; iowrite64((NTB_CTRL_LUT_EN | (sndev->self_partition << 1) | addr), &ctl->lut_entry[peer_lut_index(sndev, idx)]); } static int switchtec_ntb_mw_set_trans(struct ntb_dev *ntb, int pidx, int widx, dma_addr_t addr, resource_size_t size) { struct switchtec_ntb *sndev = ntb_sndev(ntb); struct ntb_ctrl_regs __iomem *ctl = sndev->mmio_peer_ctrl; int xlate_pos = ilog2(size); int nr_direct_mw = sndev->peer_nr_direct_mw; int rc; if (pidx != NTB_DEF_PEER_IDX) return -EINVAL; dev_dbg(&sndev->stdev->dev, "MW %d: part %d addr %pad size %pap\n", widx, pidx, &addr, &size); if (widx >= switchtec_ntb_mw_count(ntb, pidx)) return -EINVAL; if (size != 0 && xlate_pos < 12) return -EINVAL; if (!IS_ALIGNED(addr, BIT_ULL(xlate_pos))) { /* * In certain circumstances we can get a buffer that is * not aligned to its size. (Most of the time * dma_alloc_coherent ensures this). This can happen when * using large buffers allocated by the CMA * (see CMA_CONFIG_ALIGNMENT) */ dev_err(&sndev->stdev->dev, "ERROR: Memory window address is not aligned to its size!\n"); return -EINVAL; } rc = switchtec_ntb_part_op(sndev, ctl, NTB_CTRL_PART_OP_LOCK, NTB_CTRL_PART_STATUS_LOCKED); if (rc) return rc; if (size == 0) { if (widx < nr_direct_mw) switchtec_ntb_mw_clr_direct(sndev, widx); else switchtec_ntb_mw_clr_lut(sndev, widx); } else { if (widx < nr_direct_mw) switchtec_ntb_mw_set_direct(sndev, widx, addr, size); else switchtec_ntb_mw_set_lut(sndev, widx, addr, size); } rc = switchtec_ntb_part_op(sndev, ctl, NTB_CTRL_PART_OP_CFG, NTB_CTRL_PART_STATUS_NORMAL); if (rc == -EIO) { dev_err(&sndev->stdev->dev, "Hardware reported an error configuring mw %d: %08x\n", widx, ioread32(&ctl->bar_error)); if (widx < nr_direct_mw) switchtec_ntb_mw_clr_direct(sndev, widx); else switchtec_ntb_mw_clr_lut(sndev, widx); switchtec_ntb_part_op(sndev, ctl, NTB_CTRL_PART_OP_CFG, NTB_CTRL_PART_STATUS_NORMAL); } return rc; } static int switchtec_ntb_peer_mw_count(struct ntb_dev *ntb) { struct switchtec_ntb *sndev = ntb_sndev(ntb); int nr_lut_mw = sndev->nr_lut_mw - sndev->nr_rsvd_luts; return sndev->nr_direct_mw + (use_lut_mws ? nr_lut_mw : 0); } static int switchtec_ntb_direct_get_addr(struct switchtec_ntb *sndev, int idx, phys_addr_t *base, resource_size_t *size) { int bar = sndev->direct_mw_to_bar[idx]; size_t offset = 0; if (bar < 0) return -EINVAL; if (idx == 0) { /* * This is the direct BAR shared with the LUTs * which means the actual window will be offset * by the size of all the LUT entries. */ offset = LUT_SIZE * sndev->nr_lut_mw; } if (base) *base = pci_resource_start(sndev->ntb.pdev, bar) + offset; if (size) { *size = pci_resource_len(sndev->ntb.pdev, bar) - offset; if (offset && *size > offset) *size = offset; if (*size > max_mw_size) *size = max_mw_size; } return 0; } static int switchtec_ntb_lut_get_addr(struct switchtec_ntb *sndev, int idx, phys_addr_t *base, resource_size_t *size) { int bar = sndev->direct_mw_to_bar[0]; int offset; offset = LUT_SIZE * lut_index(sndev, idx); if (base) *base = pci_resource_start(sndev->ntb.pdev, bar) + offset; if (size) *size = LUT_SIZE; return 0; } static int switchtec_ntb_peer_mw_get_addr(struct ntb_dev *ntb, int idx, phys_addr_t *base, resource_size_t *size) { struct switchtec_ntb *sndev = ntb_sndev(ntb); if (idx < sndev->nr_direct_mw) return switchtec_ntb_direct_get_addr(sndev, idx, base, size); else if (idx < switchtec_ntb_peer_mw_count(ntb)) return switchtec_ntb_lut_get_addr(sndev, idx, base, size); else return -EINVAL; } static void switchtec_ntb_part_link_speed(struct switchtec_ntb *sndev, int partition, enum ntb_speed *speed, enum ntb_width *width) { struct switchtec_dev *stdev = sndev->stdev; struct part_cfg_regs __iomem *part_cfg = &stdev->mmio_part_cfg_all[partition]; u32 pff = ioread32(&part_cfg->vep_pff_inst_id) & 0xFF; u32 linksta = ioread32(&stdev->mmio_pff_csr[pff].pci_cap_region[13]); if (speed) *speed = (linksta >> 16) & 0xF; if (width) *width = (linksta >> 20) & 0x3F; } static void switchtec_ntb_set_link_speed(struct switchtec_ntb *sndev) { enum ntb_speed self_speed, peer_speed; enum ntb_width self_width, peer_width; if (!sndev->link_is_up) { sndev->link_speed = NTB_SPEED_NONE; sndev->link_width = NTB_WIDTH_NONE; return; } switchtec_ntb_part_link_speed(sndev, sndev->self_partition, &self_speed, &self_width); switchtec_ntb_part_link_speed(sndev, sndev->peer_partition, &peer_speed, &peer_width); sndev->link_speed = min(self_speed, peer_speed); sndev->link_width = min(self_width, peer_width); } static int crosslink_is_enabled(struct switchtec_ntb *sndev) { struct ntb_info_regs __iomem *inf = sndev->mmio_ntb; return ioread8(&inf->ntp_info[sndev->peer_partition].xlink_enabled); } static void crosslink_init_dbmsgs(struct switchtec_ntb *sndev) { int i; u32 msg_map = 0; if (!crosslink_is_enabled(sndev)) return; for (i = 0; i < ARRAY_SIZE(sndev->mmio_peer_dbmsg->imsg); i++) { int m = i | sndev->self_partition << 2; msg_map |= m << i * 8; } iowrite32(msg_map, &sndev->mmio_peer_dbmsg->msg_map); iowrite64(sndev->db_valid_mask << sndev->db_peer_shift, &sndev->mmio_peer_dbmsg->odb_mask); } enum switchtec_msg { LINK_MESSAGE = 0, MSG_LINK_UP = 1, MSG_LINK_DOWN = 2, MSG_CHECK_LINK = 3, MSG_LINK_FORCE_DOWN = 4, }; static int switchtec_ntb_reinit_peer(struct switchtec_ntb *sndev); static void switchtec_ntb_link_status_update(struct switchtec_ntb *sndev) { int link_sta; int old = sndev->link_is_up; link_sta = sndev->self_shared->link_sta; if (link_sta) { u64 peer = ioread64(&sndev->peer_shared->magic); if ((peer & 0xFFFFFFFF) == SWITCHTEC_NTB_MAGIC) link_sta = peer >> 32; else link_sta = 0; } sndev->link_is_up = link_sta; switchtec_ntb_set_link_speed(sndev); if (link_sta != old) { switchtec_ntb_send_msg(sndev, LINK_MESSAGE, MSG_CHECK_LINK); ntb_link_event(&sndev->ntb); dev_info(&sndev->stdev->dev, "ntb link %s\n", link_sta ? "up" : "down"); if (link_sta) crosslink_init_dbmsgs(sndev); } } static void check_link_status_work(struct work_struct *work) { struct switchtec_ntb *sndev; sndev = container_of(work, struct switchtec_ntb, check_link_status_work); if (sndev->link_force_down) { sndev->link_force_down = false; switchtec_ntb_reinit_peer(sndev); if (sndev->link_is_up) { sndev->link_is_up = 0; ntb_link_event(&sndev->ntb); dev_info(&sndev->stdev->dev, "ntb link forced down\n"); } return; } switchtec_ntb_link_status_update(sndev); } static void switchtec_ntb_check_link(struct switchtec_ntb *sndev, enum switchtec_msg msg) { if (msg == MSG_LINK_FORCE_DOWN) sndev->link_force_down = true; schedule_work(&sndev->check_link_status_work); } static void switchtec_ntb_link_notification(struct switchtec_dev *stdev) { struct switchtec_ntb *sndev = stdev->sndev; switchtec_ntb_check_link(sndev, MSG_CHECK_LINK); } static u64 switchtec_ntb_link_is_up(struct ntb_dev *ntb, enum ntb_speed *speed, enum ntb_width *width) { struct switchtec_ntb *sndev = ntb_sndev(ntb); if (speed) *speed = sndev->link_speed; if (width) *width = sndev->link_width; return sndev->link_is_up; } static int switchtec_ntb_link_enable(struct ntb_dev *ntb, enum ntb_speed max_speed, enum ntb_width max_width) { struct switchtec_ntb *sndev = ntb_sndev(ntb); dev_dbg(&sndev->stdev->dev, "enabling link\n"); sndev->self_shared->link_sta = 1; switchtec_ntb_send_msg(sndev, LINK_MESSAGE, MSG_LINK_UP); switchtec_ntb_link_status_update(sndev); return 0; } static int switchtec_ntb_link_disable(struct ntb_dev *ntb) { struct switchtec_ntb *sndev = ntb_sndev(ntb); dev_dbg(&sndev->stdev->dev, "disabling link\n"); sndev->self_shared->link_sta = 0; switchtec_ntb_send_msg(sndev, LINK_MESSAGE, MSG_LINK_DOWN); switchtec_ntb_link_status_update(sndev); return 0; } static u64 switchtec_ntb_db_valid_mask(struct ntb_dev *ntb) { struct switchtec_ntb *sndev = ntb_sndev(ntb); return sndev->db_valid_mask; } static int switchtec_ntb_db_vector_count(struct ntb_dev *ntb) { return 1; } static u64 switchtec_ntb_db_vector_mask(struct ntb_dev *ntb, int db_vector) { struct switchtec_ntb *sndev = ntb_sndev(ntb); if (db_vector < 0 || db_vector > 1) return 0; return sndev->db_valid_mask; } static u64 switchtec_ntb_db_read(struct ntb_dev *ntb) { u64 ret; struct switchtec_ntb *sndev = ntb_sndev(ntb); ret = ioread64(&sndev->mmio_self_dbmsg->idb) >> sndev->db_shift; return ret & sndev->db_valid_mask; } static int switchtec_ntb_db_clear(struct ntb_dev *ntb, u64 db_bits) { struct switchtec_ntb *sndev = ntb_sndev(ntb); iowrite64(db_bits << sndev->db_shift, &sndev->mmio_self_dbmsg->idb); return 0; } static int switchtec_ntb_db_set_mask(struct ntb_dev *ntb, u64 db_bits) { unsigned long irqflags; struct switchtec_ntb *sndev = ntb_sndev(ntb); if (db_bits & ~sndev->db_valid_mask) return -EINVAL; spin_lock_irqsave(&sndev->db_mask_lock, irqflags); sndev->db_mask |= db_bits << sndev->db_shift; iowrite64(~sndev->db_mask, &sndev->mmio_self_dbmsg->idb_mask); spin_unlock_irqrestore(&sndev->db_mask_lock, irqflags); return 0; } static int switchtec_ntb_db_clear_mask(struct ntb_dev *ntb, u64 db_bits) { unsigned long irqflags; struct switchtec_ntb *sndev = ntb_sndev(ntb); if (db_bits & ~sndev->db_valid_mask) return -EINVAL; spin_lock_irqsave(&sndev->db_mask_lock, irqflags); sndev->db_mask &= ~(db_bits << sndev->db_shift); iowrite64(~sndev->db_mask, &sndev->mmio_self_dbmsg->idb_mask); spin_unlock_irqrestore(&sndev->db_mask_lock, irqflags); return 0; } static u64 switchtec_ntb_db_read_mask(struct ntb_dev *ntb) { struct switchtec_ntb *sndev = ntb_sndev(ntb); return (sndev->db_mask >> sndev->db_shift) & sndev->db_valid_mask; } static int switchtec_ntb_peer_db_addr(struct ntb_dev *ntb, phys_addr_t *db_addr, resource_size_t *db_size, u64 *db_data, int db_bit) { struct switchtec_ntb *sndev = ntb_sndev(ntb); unsigned long offset; if (unlikely(db_bit >= BITS_PER_LONG_LONG)) return -EINVAL; offset = (unsigned long)sndev->mmio_peer_dbmsg->odb - (unsigned long)sndev->stdev->mmio; offset += sndev->db_shift / 8; if (db_addr) *db_addr = pci_resource_start(ntb->pdev, 0) + offset; if (db_size) *db_size = sizeof(u32); if (db_data) *db_data = BIT_ULL(db_bit) << sndev->db_peer_shift; return 0; } static int switchtec_ntb_peer_db_set(struct ntb_dev *ntb, u64 db_bits) { struct switchtec_ntb *sndev = ntb_sndev(ntb); iowrite64(db_bits << sndev->db_peer_shift, &sndev->mmio_peer_dbmsg->odb); return 0; } static int switchtec_ntb_spad_count(struct ntb_dev *ntb) { struct switchtec_ntb *sndev = ntb_sndev(ntb); return ARRAY_SIZE(sndev->self_shared->spad); } static u32 switchtec_ntb_spad_read(struct ntb_dev *ntb, int idx) { struct switchtec_ntb *sndev = ntb_sndev(ntb); if (idx < 0 || idx >= ARRAY_SIZE(sndev->self_shared->spad)) return 0; if (!sndev->self_shared) return 0; return sndev->self_shared->spad[idx]; } static int switchtec_ntb_spad_write(struct ntb_dev *ntb, int idx, u32 val) { struct switchtec_ntb *sndev = ntb_sndev(ntb); if (idx < 0 || idx >= ARRAY_SIZE(sndev->self_shared->spad)) return -EINVAL; if (!sndev->self_shared) return -EIO; sndev->self_shared->spad[idx] = val; return 0; } static u32 switchtec_ntb_peer_spad_read(struct ntb_dev *ntb, int pidx, int sidx) { struct switchtec_ntb *sndev = ntb_sndev(ntb); if (pidx != NTB_DEF_PEER_IDX) return -EINVAL; if (sidx < 0 || sidx >= ARRAY_SIZE(sndev->peer_shared->spad)) return 0; if (!sndev->peer_shared) return 0; return ioread32(&sndev->peer_shared->spad[sidx]); } static int switchtec_ntb_peer_spad_write(struct ntb_dev *ntb, int pidx, int sidx, u32 val) { struct switchtec_ntb *sndev = ntb_sndev(ntb); if (pidx != NTB_DEF_PEER_IDX) return -EINVAL; if (sidx < 0 || sidx >= ARRAY_SIZE(sndev->peer_shared->spad)) return -EINVAL; if (!sndev->peer_shared) return -EIO; iowrite32(val, &sndev->peer_shared->spad[sidx]); return 0; } static int switchtec_ntb_peer_spad_addr(struct ntb_dev *ntb, int pidx, int sidx, phys_addr_t *spad_addr) { struct switchtec_ntb *sndev = ntb_sndev(ntb); unsigned long offset; if (pidx != NTB_DEF_PEER_IDX) return -EINVAL; offset = (unsigned long)&sndev->peer_shared->spad[sidx] - (unsigned long)sndev->stdev->mmio; if (spad_addr) *spad_addr = pci_resource_start(ntb->pdev, 0) + offset; return 0; } static const struct ntb_dev_ops switchtec_ntb_ops = { .mw_count = switchtec_ntb_mw_count, .mw_get_align = switchtec_ntb_mw_get_align, .mw_set_trans = switchtec_ntb_mw_set_trans, .peer_mw_count = switchtec_ntb_peer_mw_count, .peer_mw_get_addr = switchtec_ntb_peer_mw_get_addr, .link_is_up = switchtec_ntb_link_is_up, .link_enable = switchtec_ntb_link_enable, .link_disable = switchtec_ntb_link_disable, .db_valid_mask = switchtec_ntb_db_valid_mask, .db_vector_count = switchtec_ntb_db_vector_count, .db_vector_mask = switchtec_ntb_db_vector_mask, .db_read = switchtec_ntb_db_read, .db_clear = switchtec_ntb_db_clear, .db_set_mask = switchtec_ntb_db_set_mask, .db_clear_mask = switchtec_ntb_db_clear_mask, .db_read_mask = switchtec_ntb_db_read_mask, .peer_db_addr = switchtec_ntb_peer_db_addr, .peer_db_set = switchtec_ntb_peer_db_set, .spad_count = switchtec_ntb_spad_count, .spad_read = switchtec_ntb_spad_read, .spad_write = switchtec_ntb_spad_write, .peer_spad_read = switchtec_ntb_peer_spad_read, .peer_spad_write = switchtec_ntb_peer_spad_write, .peer_spad_addr = switchtec_ntb_peer_spad_addr, }; static int switchtec_ntb_init_sndev(struct switchtec_ntb *sndev) { u64 tpart_vec; int self; u64 part_map; sndev->ntb.pdev = sndev->stdev->pdev; sndev->ntb.topo = NTB_TOPO_SWITCH; sndev->ntb.ops = &switchtec_ntb_ops; INIT_WORK(&sndev->check_link_status_work, check_link_status_work); sndev->link_force_down = false; sndev->self_partition = sndev->stdev->partition; sndev->mmio_ntb = sndev->stdev->mmio_ntb; self = sndev->self_partition; tpart_vec = ioread32(&sndev->mmio_ntb->ntp_info[self].target_part_high); tpart_vec <<= 32; tpart_vec |= ioread32(&sndev->mmio_ntb->ntp_info[self].target_part_low); part_map = ioread64(&sndev->mmio_ntb->ep_map); tpart_vec &= part_map; part_map &= ~(1 << sndev->self_partition); if (!tpart_vec) { if (sndev->stdev->partition_count != 2) { dev_err(&sndev->stdev->dev, "ntb target partition not defined\n"); return -ENODEV; } if (!part_map) { dev_err(&sndev->stdev->dev, "peer partition is not NT partition\n"); return -ENODEV; } sndev->peer_partition = __ffs64(part_map); } else { if (__ffs64(tpart_vec) != (fls64(tpart_vec) - 1)) { dev_err(&sndev->stdev->dev, "ntb driver only supports 1 pair of 1-1 ntb mapping\n"); return -ENODEV; } sndev->peer_partition = __ffs64(tpart_vec); if (!(part_map & (1ULL << sndev->peer_partition))) { dev_err(&sndev->stdev->dev, "ntb target partition is not NT partition\n"); return -ENODEV; } } dev_dbg(&sndev->stdev->dev, "Partition ID %d of %d\n", sndev->self_partition, sndev->stdev->partition_count); sndev->mmio_ctrl = (void * __iomem)sndev->mmio_ntb + SWITCHTEC_NTB_REG_CTRL_OFFSET; sndev->mmio_dbmsg = (void * __iomem)sndev->mmio_ntb + SWITCHTEC_NTB_REG_DBMSG_OFFSET; sndev->mmio_self_ctrl = &sndev->mmio_ctrl[sndev->self_partition]; sndev->mmio_peer_ctrl = &sndev->mmio_ctrl[sndev->peer_partition]; sndev->mmio_self_dbmsg = &sndev->mmio_dbmsg[sndev->self_partition]; sndev->mmio_peer_dbmsg = sndev->mmio_self_dbmsg; return 0; } static int config_rsvd_lut_win(struct switchtec_ntb *sndev, struct ntb_ctrl_regs __iomem *ctl, int lut_idx, int partition, u64 addr) { int peer_bar = sndev->peer_direct_mw_to_bar[0]; u32 ctl_val; int rc; rc = switchtec_ntb_part_op(sndev, ctl, NTB_CTRL_PART_OP_LOCK, NTB_CTRL_PART_STATUS_LOCKED); if (rc) return rc; ctl_val = ioread32(&ctl->bar_entry[peer_bar].ctl); ctl_val &= 0xFF; ctl_val |= NTB_CTRL_BAR_LUT_WIN_EN; ctl_val |= ilog2(LUT_SIZE) << 8; ctl_val |= (sndev->nr_lut_mw - 1) << 14; iowrite32(ctl_val, &ctl->bar_entry[peer_bar].ctl); iowrite64((NTB_CTRL_LUT_EN | (partition << 1) | addr), &ctl->lut_entry[lut_idx]); rc = switchtec_ntb_part_op(sndev, ctl, NTB_CTRL_PART_OP_CFG, NTB_CTRL_PART_STATUS_NORMAL); if (rc) { u32 bar_error, lut_error; bar_error = ioread32(&ctl->bar_error); lut_error = ioread32(&ctl->lut_error); dev_err(&sndev->stdev->dev, "Error setting up reserved lut window: %08x / %08x\n", bar_error, lut_error); return rc; } return 0; } static int config_req_id_table(struct switchtec_ntb *sndev, struct ntb_ctrl_regs __iomem *mmio_ctrl, int *req_ids, int count) { int i, rc = 0; u32 error; u32 proxy_id; if (ioread16(&mmio_ctrl->req_id_table_size) < count) { dev_err(&sndev->stdev->dev, "Not enough requester IDs available.\n"); return -EFAULT; } rc = switchtec_ntb_part_op(sndev, mmio_ctrl, NTB_CTRL_PART_OP_LOCK, NTB_CTRL_PART_STATUS_LOCKED); if (rc) return rc; for (i = 0; i < count; i++) { iowrite32(req_ids[i] << 16 | NTB_CTRL_REQ_ID_EN, &mmio_ctrl->req_id_table[i]); proxy_id = ioread32(&mmio_ctrl->req_id_table[i]); dev_dbg(&sndev->stdev->dev, "Requester ID %02X:%02X.%X -> BB:%02X.%X\n", req_ids[i] >> 8, (req_ids[i] >> 3) & 0x1F, req_ids[i] & 0x7, (proxy_id >> 4) & 0x1F, (proxy_id >> 1) & 0x7); } rc = switchtec_ntb_part_op(sndev, mmio_ctrl, NTB_CTRL_PART_OP_CFG, NTB_CTRL_PART_STATUS_NORMAL); if (rc == -EIO) { error = ioread32(&mmio_ctrl->req_id_error); dev_err(&sndev->stdev->dev, "Error setting up the requester ID table: %08x\n", error); } return 0; } static int crosslink_setup_mws(struct switchtec_ntb *sndev, int ntb_lut_idx, u64 *mw_addrs, int mw_count) { int rc, i; struct ntb_ctrl_regs __iomem *ctl = sndev->mmio_self_ctrl; u64 addr; size_t size, offset; int bar; int xlate_pos; u32 ctl_val; rc = switchtec_ntb_part_op(sndev, ctl, NTB_CTRL_PART_OP_LOCK, NTB_CTRL_PART_STATUS_LOCKED); if (rc) return rc; for (i = 0; i < sndev->nr_lut_mw; i++) { if (i == ntb_lut_idx) continue; addr = mw_addrs[0] + LUT_SIZE * i; iowrite64((NTB_CTRL_LUT_EN | (sndev->peer_partition << 1) | addr), &ctl->lut_entry[i]); } sndev->nr_direct_mw = min_t(int, sndev->nr_direct_mw, mw_count); for (i = 0; i < sndev->nr_direct_mw; i++) { bar = sndev->direct_mw_to_bar[i]; offset = (i == 0) ? LUT_SIZE * sndev->nr_lut_mw : 0; addr = mw_addrs[i] + offset; size = pci_resource_len(sndev->ntb.pdev, bar) - offset; xlate_pos = ilog2(size); if (offset && size > offset) size = offset; ctl_val = ioread32(&ctl->bar_entry[bar].ctl); ctl_val |= NTB_CTRL_BAR_DIR_WIN_EN; iowrite32(ctl_val, &ctl->bar_entry[bar].ctl); iowrite32(xlate_pos | (lower_32_bits(size) & 0xFFFFF000), &ctl->bar_entry[bar].win_size); iowrite32(upper_32_bits(size), &ctl->bar_ext_entry[bar].win_size); iowrite64(sndev->peer_partition | addr, &ctl->bar_entry[bar].xlate_addr); } rc = switchtec_ntb_part_op(sndev, ctl, NTB_CTRL_PART_OP_CFG, NTB_CTRL_PART_STATUS_NORMAL); if (rc) { u32 bar_error, lut_error; bar_error = ioread32(&ctl->bar_error); lut_error = ioread32(&ctl->lut_error); dev_err(&sndev->stdev->dev, "Error setting up cross link windows: %08x / %08x\n", bar_error, lut_error); return rc; } return 0; } static int crosslink_setup_req_ids(struct switchtec_ntb *sndev, struct ntb_ctrl_regs __iomem *mmio_ctrl) { int req_ids[16]; int i; u32 proxy_id; for (i = 0; i < ARRAY_SIZE(req_ids); i++) { proxy_id = ioread32(&sndev->mmio_self_ctrl->req_id_table[i]); if (!(proxy_id & NTB_CTRL_REQ_ID_EN)) break; req_ids[i] = ((proxy_id >> 1) & 0xFF); } return config_req_id_table(sndev, mmio_ctrl, req_ids, i); } /* * In crosslink configuration there is a virtual partition in the * middle of the two switches. The BARs in this partition have to be * enumerated and assigned addresses. */ static int crosslink_enum_partition(struct switchtec_ntb *sndev, u64 *bar_addrs) { struct part_cfg_regs __iomem *part_cfg = &sndev->stdev->mmio_part_cfg_all[sndev->peer_partition]; u32 pff = ioread32(&part_cfg->vep_pff_inst_id) & 0xFF; struct pff_csr_regs __iomem *mmio_pff = &sndev->stdev->mmio_pff_csr[pff]; const u64 bar_space = 0x1000000000LL; u64 bar_addr; int bar_cnt = 0; int i; iowrite16(0x6, &mmio_pff->pcicmd); for (i = 0; i < ARRAY_SIZE(mmio_pff->pci_bar64); i++) { iowrite64(bar_space * i, &mmio_pff->pci_bar64[i]); bar_addr = ioread64(&mmio_pff->pci_bar64[i]); bar_addr &= ~0xf; dev_dbg(&sndev->stdev->dev, "Crosslink BAR%d addr: %llx\n", i*2, bar_addr); if (bar_addr != bar_space * i) continue; bar_addrs[bar_cnt++] = bar_addr; } return bar_cnt; } static int switchtec_ntb_init_crosslink(struct switchtec_ntb *sndev) { int rc; int bar = sndev->direct_mw_to_bar[0]; const int ntb_lut_idx = 1; u64 bar_addrs[6]; u64 addr; int offset; int bar_cnt; if (!crosslink_is_enabled(sndev)) return 0; dev_info(&sndev->stdev->dev, "Using crosslink configuration\n"); sndev->ntb.topo = NTB_TOPO_CROSSLINK; bar_cnt = crosslink_enum_partition(sndev, bar_addrs); if (bar_cnt < sndev->nr_direct_mw + 1) { dev_err(&sndev->stdev->dev, "Error enumerating crosslink partition\n"); return -EINVAL; } addr = (bar_addrs[0] + SWITCHTEC_GAS_NTB_OFFSET + SWITCHTEC_NTB_REG_DBMSG_OFFSET + sizeof(struct ntb_dbmsg_regs) * sndev->peer_partition); offset = addr & (LUT_SIZE - 1); addr -= offset; rc = config_rsvd_lut_win(sndev, sndev->mmio_self_ctrl, ntb_lut_idx, sndev->peer_partition, addr); if (rc) return rc; rc = crosslink_setup_mws(sndev, ntb_lut_idx, &bar_addrs[1], bar_cnt - 1); if (rc) return rc; rc = crosslink_setup_req_ids(sndev, sndev->mmio_peer_ctrl); if (rc) return rc; sndev->mmio_xlink_win = pci_iomap_range(sndev->stdev->pdev, bar, LUT_SIZE, LUT_SIZE); if (!sndev->mmio_xlink_win) { rc = -ENOMEM; return rc; } sndev->mmio_peer_dbmsg = sndev->mmio_xlink_win + offset; sndev->nr_rsvd_luts++; crosslink_init_dbmsgs(sndev); return 0; } static void switchtec_ntb_deinit_crosslink(struct switchtec_ntb *sndev) { if (sndev->mmio_xlink_win) pci_iounmap(sndev->stdev->pdev, sndev->mmio_xlink_win); } static int map_bars(int *map, struct ntb_ctrl_regs __iomem *ctrl) { int i; int cnt = 0; for (i = 0; i < ARRAY_SIZE(ctrl->bar_entry); i++) { u32 r = ioread32(&ctrl->bar_entry[i].ctl); if (r & NTB_CTRL_BAR_VALID) map[cnt++] = i; } return cnt; } static void switchtec_ntb_init_mw(struct switchtec_ntb *sndev) { sndev->nr_direct_mw = map_bars(sndev->direct_mw_to_bar, sndev->mmio_self_ctrl); sndev->nr_lut_mw = ioread16(&sndev->mmio_self_ctrl->lut_table_entries); sndev->nr_lut_mw = rounddown_pow_of_two(sndev->nr_lut_mw); dev_dbg(&sndev->stdev->dev, "MWs: %d direct, %d lut\n", sndev->nr_direct_mw, sndev->nr_lut_mw); sndev->peer_nr_direct_mw = map_bars(sndev->peer_direct_mw_to_bar, sndev->mmio_peer_ctrl); sndev->peer_nr_lut_mw = ioread16(&sndev->mmio_peer_ctrl->lut_table_entries); sndev->peer_nr_lut_mw = rounddown_pow_of_two(sndev->peer_nr_lut_mw); dev_dbg(&sndev->stdev->dev, "Peer MWs: %d direct, %d lut\n", sndev->peer_nr_direct_mw, sndev->peer_nr_lut_mw); } /* * There are 64 doorbells in the switch hardware but this is * shared among all partitions. So we must split them in half * (32 for each partition). However, the message interrupts are * also shared with the top 4 doorbells so we just limit this to * 28 doorbells per partition. * * In crosslink mode, each side has it's own dbmsg register so * they can each use all 60 of the available doorbells. */ static void switchtec_ntb_init_db(struct switchtec_ntb *sndev) { sndev->db_mask = 0x0FFFFFFFFFFFFFFFULL; if (sndev->mmio_peer_dbmsg != sndev->mmio_self_dbmsg) { sndev->db_shift = 0; sndev->db_peer_shift = 0; sndev->db_valid_mask = sndev->db_mask; } else if (sndev->self_partition < sndev->peer_partition) { sndev->db_shift = 0; sndev->db_peer_shift = 32; sndev->db_valid_mask = 0x0FFFFFFF; } else { sndev->db_shift = 32; sndev->db_peer_shift = 0; sndev->db_valid_mask = 0x0FFFFFFF; } iowrite64(~sndev->db_mask, &sndev->mmio_self_dbmsg->idb_mask); iowrite64(sndev->db_valid_mask << sndev->db_peer_shift, &sndev->mmio_peer_dbmsg->odb_mask); dev_dbg(&sndev->stdev->dev, "dbs: shift %d/%d, mask %016llx\n", sndev->db_shift, sndev->db_peer_shift, sndev->db_valid_mask); } static void switchtec_ntb_init_msgs(struct switchtec_ntb *sndev) { int i; u32 msg_map = 0; for (i = 0; i < ARRAY_SIZE(sndev->mmio_self_dbmsg->imsg); i++) { int m = i | sndev->peer_partition << 2; msg_map |= m << i * 8; } iowrite32(msg_map, &sndev->mmio_self_dbmsg->msg_map); for (i = 0; i < ARRAY_SIZE(sndev->mmio_self_dbmsg->imsg); i++) iowrite64(NTB_DBMSG_IMSG_STATUS | NTB_DBMSG_IMSG_MASK, &sndev->mmio_self_dbmsg->imsg[i]); } static int switchtec_ntb_init_req_id_table(struct switchtec_ntb *sndev) { int req_ids[2]; /* * Root Complex Requester ID (which is 0:00.0) */ req_ids[0] = 0; /* * Host Bridge Requester ID (as read from the mmap address) */ req_ids[1] = ioread16(&sndev->mmio_ntb->requester_id); return config_req_id_table(sndev, sndev->mmio_self_ctrl, req_ids, ARRAY_SIZE(req_ids)); } static void switchtec_ntb_init_shared(struct switchtec_ntb *sndev) { int i; memset(sndev->self_shared, 0, LUT_SIZE); sndev->self_shared->magic = SWITCHTEC_NTB_MAGIC; sndev->self_shared->partition_id = sndev->stdev->partition; for (i = 0; i < sndev->nr_direct_mw; i++) { int bar = sndev->direct_mw_to_bar[i]; resource_size_t sz = pci_resource_len(sndev->stdev->pdev, bar); if (i == 0) sz = min_t(resource_size_t, sz, LUT_SIZE * sndev->nr_lut_mw); sndev->self_shared->mw_sizes[i] = sz; } for (i = 0; i < sndev->nr_lut_mw; i++) { int idx = sndev->nr_direct_mw + i; sndev->self_shared->mw_sizes[idx] = LUT_SIZE; } } static int switchtec_ntb_init_shared_mw(struct switchtec_ntb *sndev) { int self_bar = sndev->direct_mw_to_bar[0]; int rc; sndev->nr_rsvd_luts++; sndev->self_shared = dma_alloc_coherent(&sndev->stdev->pdev->dev, LUT_SIZE, &sndev->self_shared_dma, GFP_KERNEL); if (!sndev->self_shared) { dev_err(&sndev->stdev->dev, "unable to allocate memory for shared mw\n"); return -ENOMEM; } switchtec_ntb_init_shared(sndev); rc = config_rsvd_lut_win(sndev, sndev->mmio_peer_ctrl, 0, sndev->self_partition, sndev->self_shared_dma); if (rc) goto unalloc_and_exit; sndev->peer_shared = pci_iomap(sndev->stdev->pdev, self_bar, LUT_SIZE); if (!sndev->peer_shared) { rc = -ENOMEM; goto unalloc_and_exit; } dev_dbg(&sndev->stdev->dev, "Shared MW Ready\n"); return 0; unalloc_and_exit: dma_free_coherent(&sndev->stdev->pdev->dev, LUT_SIZE, sndev->self_shared, sndev->self_shared_dma); return rc; } static void switchtec_ntb_deinit_shared_mw(struct switchtec_ntb *sndev) { if (sndev->peer_shared) pci_iounmap(sndev->stdev->pdev, sndev->peer_shared); if (sndev->self_shared) dma_free_coherent(&sndev->stdev->pdev->dev, LUT_SIZE, sndev->self_shared, sndev->self_shared_dma); sndev->nr_rsvd_luts--; } static irqreturn_t switchtec_ntb_doorbell_isr(int irq, void *dev) { struct switchtec_ntb *sndev = dev; dev_dbg(&sndev->stdev->dev, "doorbell\n"); ntb_db_event(&sndev->ntb, 0); return IRQ_HANDLED; } static irqreturn_t switchtec_ntb_message_isr(int irq, void *dev) { int i; struct switchtec_ntb *sndev = dev; for (i = 0; i < ARRAY_SIZE(sndev->mmio_self_dbmsg->imsg); i++) { u64 msg = ioread64(&sndev->mmio_self_dbmsg->imsg[i]); if (msg & NTB_DBMSG_IMSG_STATUS) { dev_dbg(&sndev->stdev->dev, "message: %d %08x\n", i, (u32)msg); iowrite8(1, &sndev->mmio_self_dbmsg->imsg[i].status); if (i == LINK_MESSAGE) switchtec_ntb_check_link(sndev, msg); } } return IRQ_HANDLED; } static int switchtec_ntb_init_db_msg_irq(struct switchtec_ntb *sndev) { int i; int rc; int doorbell_irq = 0; int message_irq = 0; int event_irq; int idb_vecs = sizeof(sndev->mmio_self_dbmsg->idb_vec_map); event_irq = ioread32(&sndev->stdev->mmio_part_cfg->vep_vector_number); while (doorbell_irq == event_irq) doorbell_irq++; while (message_irq == doorbell_irq || message_irq == event_irq) message_irq++; dev_dbg(&sndev->stdev->dev, "irqs - event: %d, db: %d, msgs: %d\n", event_irq, doorbell_irq, message_irq); for (i = 0; i < idb_vecs - 4; i++) iowrite8(doorbell_irq, &sndev->mmio_self_dbmsg->idb_vec_map[i]); for (; i < idb_vecs; i++) iowrite8(message_irq, &sndev->mmio_self_dbmsg->idb_vec_map[i]); sndev->doorbell_irq = pci_irq_vector(sndev->stdev->pdev, doorbell_irq); sndev->message_irq = pci_irq_vector(sndev->stdev->pdev, message_irq); rc = request_irq(sndev->doorbell_irq, switchtec_ntb_doorbell_isr, 0, "switchtec_ntb_doorbell", sndev); if (rc) return rc; rc = request_irq(sndev->message_irq, switchtec_ntb_message_isr, 0, "switchtec_ntb_message", sndev); if (rc) { free_irq(sndev->doorbell_irq, sndev); return rc; } return 0; } static void switchtec_ntb_deinit_db_msg_irq(struct switchtec_ntb *sndev) { free_irq(sndev->doorbell_irq, sndev); free_irq(sndev->message_irq, sndev); } static int switchtec_ntb_reinit_peer(struct switchtec_ntb *sndev) { int rc; if (crosslink_is_enabled(sndev)) return 0; dev_info(&sndev->stdev->dev, "reinitialize shared memory window\n"); rc = config_rsvd_lut_win(sndev, sndev->mmio_peer_ctrl, 0, sndev->self_partition, sndev->self_shared_dma); return rc; } static int switchtec_ntb_add(struct device *dev) { struct switchtec_dev *stdev = to_stdev(dev); struct switchtec_ntb *sndev; int rc; stdev->sndev = NULL; if (stdev->pdev->class != (PCI_CLASS_BRIDGE_OTHER << 8)) return -ENODEV; sndev = kzalloc_node(sizeof(*sndev), GFP_KERNEL, dev_to_node(dev)); if (!sndev) return -ENOMEM; sndev->stdev = stdev; rc = switchtec_ntb_init_sndev(sndev); if (rc) goto free_and_exit; switchtec_ntb_init_mw(sndev); rc = switchtec_ntb_init_req_id_table(sndev); if (rc) goto free_and_exit; rc = switchtec_ntb_init_crosslink(sndev); if (rc) goto free_and_exit; switchtec_ntb_init_db(sndev); switchtec_ntb_init_msgs(sndev); rc = switchtec_ntb_init_shared_mw(sndev); if (rc) goto deinit_crosslink; rc = switchtec_ntb_init_db_msg_irq(sndev); if (rc) goto deinit_shared_and_exit; /* * If this host crashed, the other host may think the link is * still up. Tell them to force it down (it will go back up * once we register the ntb device). */ switchtec_ntb_send_msg(sndev, LINK_MESSAGE, MSG_LINK_FORCE_DOWN); rc = ntb_register_device(&sndev->ntb); if (rc) goto deinit_and_exit; stdev->sndev = sndev; stdev->link_notifier = switchtec_ntb_link_notification; dev_info(dev, "NTB device registered\n"); return 0; deinit_and_exit: switchtec_ntb_deinit_db_msg_irq(sndev); deinit_shared_and_exit: switchtec_ntb_deinit_shared_mw(sndev); deinit_crosslink: switchtec_ntb_deinit_crosslink(sndev); free_and_exit: kfree(sndev); dev_err(dev, "failed to register ntb device: %d\n", rc); return rc; } static void switchtec_ntb_remove(struct device *dev) { struct switchtec_dev *stdev = to_stdev(dev); struct switchtec_ntb *sndev = stdev->sndev; if (!sndev) return; stdev->link_notifier = NULL; stdev->sndev = NULL; ntb_unregister_device(&sndev->ntb); switchtec_ntb_deinit_db_msg_irq(sndev); switchtec_ntb_deinit_shared_mw(sndev); switchtec_ntb_deinit_crosslink(sndev); cancel_work_sync(&sndev->check_link_status_work); kfree(sndev); dev_info(dev, "ntb device unregistered\n"); } static struct class_interface switchtec_interface = { .add_dev = switchtec_ntb_add, .remove_dev = switchtec_ntb_remove, }; static int __init switchtec_ntb_init(void) { switchtec_interface.class = &switchtec_class; return class_interface_register(&switchtec_interface); } module_init(switchtec_ntb_init); static void __exit switchtec_ntb_exit(void) { class_interface_unregister(&switchtec_interface); } module_exit(switchtec_ntb_exit);