818 lines
20 KiB
C
818 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Copyright(c) 2017 - 2019 Pensando Systems, Inc */
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/types.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/io.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/etherdevice.h>
|
|
#include "ionic.h"
|
|
#include "ionic_dev.h"
|
|
#include "ionic_lif.h"
|
|
|
|
static void ionic_watchdog_cb(struct timer_list *t)
|
|
{
|
|
struct ionic *ionic = from_timer(ionic, t, watchdog_timer);
|
|
struct ionic_lif *lif = ionic->lif;
|
|
struct ionic_deferred_work *work;
|
|
int hb;
|
|
|
|
mod_timer(&ionic->watchdog_timer,
|
|
round_jiffies(jiffies + ionic->watchdog_period));
|
|
|
|
if (!lif)
|
|
return;
|
|
|
|
hb = ionic_heartbeat_check(ionic);
|
|
dev_dbg(ionic->dev, "%s: hb %d running %d UP %d\n",
|
|
__func__, hb, netif_running(lif->netdev),
|
|
test_bit(IONIC_LIF_F_UP, lif->state));
|
|
|
|
if (hb >= 0 &&
|
|
!test_bit(IONIC_LIF_F_FW_RESET, lif->state))
|
|
ionic_link_status_check_request(lif, CAN_NOT_SLEEP);
|
|
|
|
if (test_bit(IONIC_LIF_F_FILTER_SYNC_NEEDED, lif->state) &&
|
|
!test_bit(IONIC_LIF_F_FW_RESET, lif->state)) {
|
|
work = kzalloc(sizeof(*work), GFP_ATOMIC);
|
|
if (!work) {
|
|
netdev_err(lif->netdev, "rxmode change dropped\n");
|
|
return;
|
|
}
|
|
|
|
work->type = IONIC_DW_TYPE_RX_MODE;
|
|
netdev_dbg(lif->netdev, "deferred: rx_mode\n");
|
|
ionic_lif_deferred_enqueue(&lif->deferred, work);
|
|
}
|
|
}
|
|
|
|
static void ionic_watchdog_init(struct ionic *ionic)
|
|
{
|
|
struct ionic_dev *idev = &ionic->idev;
|
|
|
|
timer_setup(&ionic->watchdog_timer, ionic_watchdog_cb, 0);
|
|
ionic->watchdog_period = IONIC_WATCHDOG_SECS * HZ;
|
|
|
|
/* set times to ensure the first check will proceed */
|
|
atomic_long_set(&idev->last_check_time, jiffies - 2 * HZ);
|
|
idev->last_hb_time = jiffies - 2 * ionic->watchdog_period;
|
|
/* init as ready, so no transition if the first check succeeds */
|
|
idev->last_fw_hb = 0;
|
|
idev->fw_hb_ready = true;
|
|
idev->fw_status_ready = true;
|
|
idev->fw_generation = IONIC_FW_STS_F_GENERATION &
|
|
ioread8(&idev->dev_info_regs->fw_status);
|
|
}
|
|
|
|
void ionic_init_devinfo(struct ionic *ionic)
|
|
{
|
|
struct ionic_dev *idev = &ionic->idev;
|
|
|
|
idev->dev_info.asic_type = ioread8(&idev->dev_info_regs->asic_type);
|
|
idev->dev_info.asic_rev = ioread8(&idev->dev_info_regs->asic_rev);
|
|
|
|
memcpy_fromio(idev->dev_info.fw_version,
|
|
idev->dev_info_regs->fw_version,
|
|
IONIC_DEVINFO_FWVERS_BUFLEN);
|
|
|
|
memcpy_fromio(idev->dev_info.serial_num,
|
|
idev->dev_info_regs->serial_num,
|
|
IONIC_DEVINFO_SERIAL_BUFLEN);
|
|
|
|
idev->dev_info.fw_version[IONIC_DEVINFO_FWVERS_BUFLEN] = 0;
|
|
idev->dev_info.serial_num[IONIC_DEVINFO_SERIAL_BUFLEN] = 0;
|
|
|
|
dev_dbg(ionic->dev, "fw_version %s\n", idev->dev_info.fw_version);
|
|
}
|
|
|
|
int ionic_dev_setup(struct ionic *ionic)
|
|
{
|
|
struct ionic_dev_bar *bar = ionic->bars;
|
|
unsigned int num_bars = ionic->num_bars;
|
|
struct ionic_dev *idev = &ionic->idev;
|
|
struct device *dev = ionic->dev;
|
|
int size;
|
|
u32 sig;
|
|
|
|
/* BAR0: dev_cmd and interrupts */
|
|
if (num_bars < 1) {
|
|
dev_err(dev, "No bars found, aborting\n");
|
|
return -EFAULT;
|
|
}
|
|
|
|
if (bar->len < IONIC_BAR0_SIZE) {
|
|
dev_err(dev, "Resource bar size %lu too small, aborting\n",
|
|
bar->len);
|
|
return -EFAULT;
|
|
}
|
|
|
|
idev->dev_info_regs = bar->vaddr + IONIC_BAR0_DEV_INFO_REGS_OFFSET;
|
|
idev->dev_cmd_regs = bar->vaddr + IONIC_BAR0_DEV_CMD_REGS_OFFSET;
|
|
idev->intr_status = bar->vaddr + IONIC_BAR0_INTR_STATUS_OFFSET;
|
|
idev->intr_ctrl = bar->vaddr + IONIC_BAR0_INTR_CTRL_OFFSET;
|
|
|
|
idev->hwstamp_regs = &idev->dev_info_regs->hwstamp;
|
|
|
|
sig = ioread32(&idev->dev_info_regs->signature);
|
|
if (sig != IONIC_DEV_INFO_SIGNATURE) {
|
|
dev_err(dev, "Incompatible firmware signature %x", sig);
|
|
return -EFAULT;
|
|
}
|
|
|
|
ionic_init_devinfo(ionic);
|
|
|
|
/* BAR1: doorbells */
|
|
bar++;
|
|
if (num_bars < 2) {
|
|
dev_err(dev, "Doorbell bar missing, aborting\n");
|
|
return -EFAULT;
|
|
}
|
|
|
|
ionic_watchdog_init(ionic);
|
|
|
|
idev->db_pages = bar->vaddr;
|
|
idev->phy_db_pages = bar->bus_addr;
|
|
|
|
/* BAR2: optional controller memory mapping */
|
|
bar++;
|
|
mutex_init(&idev->cmb_inuse_lock);
|
|
if (num_bars < 3 || !ionic->bars[IONIC_PCI_BAR_CMB].len) {
|
|
idev->cmb_inuse = NULL;
|
|
return 0;
|
|
}
|
|
|
|
idev->phy_cmb_pages = bar->bus_addr;
|
|
idev->cmb_npages = bar->len / PAGE_SIZE;
|
|
size = BITS_TO_LONGS(idev->cmb_npages) * sizeof(long);
|
|
idev->cmb_inuse = kzalloc(size, GFP_KERNEL);
|
|
if (!idev->cmb_inuse)
|
|
dev_warn(dev, "No memory for CMB, disabling\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ionic_dev_teardown(struct ionic *ionic)
|
|
{
|
|
struct ionic_dev *idev = &ionic->idev;
|
|
|
|
kfree(idev->cmb_inuse);
|
|
idev->cmb_inuse = NULL;
|
|
idev->phy_cmb_pages = 0;
|
|
idev->cmb_npages = 0;
|
|
|
|
mutex_destroy(&idev->cmb_inuse_lock);
|
|
}
|
|
|
|
/* Devcmd Interface */
|
|
static bool __ionic_is_fw_running(struct ionic_dev *idev, u8 *status_ptr)
|
|
{
|
|
u8 fw_status;
|
|
|
|
if (!idev->dev_info_regs) {
|
|
if (status_ptr)
|
|
*status_ptr = 0xff;
|
|
return false;
|
|
}
|
|
|
|
fw_status = ioread8(&idev->dev_info_regs->fw_status);
|
|
if (status_ptr)
|
|
*status_ptr = fw_status;
|
|
|
|
/* firmware is useful only if the running bit is set and
|
|
* fw_status != 0xff (bad PCI read)
|
|
*/
|
|
return (fw_status != 0xff) && (fw_status & IONIC_FW_STS_F_RUNNING);
|
|
}
|
|
|
|
bool ionic_is_fw_running(struct ionic_dev *idev)
|
|
{
|
|
return __ionic_is_fw_running(idev, NULL);
|
|
}
|
|
|
|
int ionic_heartbeat_check(struct ionic *ionic)
|
|
{
|
|
unsigned long check_time, last_check_time;
|
|
struct ionic_dev *idev = &ionic->idev;
|
|
struct ionic_lif *lif = ionic->lif;
|
|
bool fw_status_ready = true;
|
|
bool fw_hb_ready;
|
|
u8 fw_generation;
|
|
u8 fw_status;
|
|
u32 fw_hb;
|
|
|
|
/* wait a least one second before testing again */
|
|
check_time = jiffies;
|
|
last_check_time = atomic_long_read(&idev->last_check_time);
|
|
do_check_time:
|
|
if (time_before(check_time, last_check_time + HZ))
|
|
return 0;
|
|
if (!atomic_long_try_cmpxchg_relaxed(&idev->last_check_time,
|
|
&last_check_time, check_time)) {
|
|
/* if called concurrently, only the first should proceed. */
|
|
dev_dbg(ionic->dev, "%s: do_check_time again\n", __func__);
|
|
goto do_check_time;
|
|
}
|
|
|
|
/* If fw_status is not ready don't bother with the generation */
|
|
if (!__ionic_is_fw_running(idev, &fw_status)) {
|
|
fw_status_ready = false;
|
|
} else {
|
|
fw_generation = fw_status & IONIC_FW_STS_F_GENERATION;
|
|
if (idev->fw_generation != fw_generation) {
|
|
dev_info(ionic->dev, "FW generation 0x%02x -> 0x%02x\n",
|
|
idev->fw_generation, fw_generation);
|
|
|
|
idev->fw_generation = fw_generation;
|
|
|
|
/* If the generation changed, the fw status is not
|
|
* ready so we need to trigger a fw-down cycle. After
|
|
* the down, the next watchdog will see the fw is up
|
|
* and the generation value stable, so will trigger
|
|
* the fw-up activity.
|
|
*
|
|
* If we had already moved to FW_RESET from a RESET event,
|
|
* it is possible that we never saw the fw_status go to 0,
|
|
* so we fake the current idev->fw_status_ready here to
|
|
* force the transition and get FW up again.
|
|
*/
|
|
if (test_bit(IONIC_LIF_F_FW_RESET, lif->state))
|
|
idev->fw_status_ready = false; /* go to running */
|
|
else
|
|
fw_status_ready = false; /* go to down */
|
|
}
|
|
}
|
|
|
|
dev_dbg(ionic->dev, "fw_status 0x%02x ready %d idev->ready %d last_hb 0x%x state 0x%02lx\n",
|
|
fw_status, fw_status_ready, idev->fw_status_ready,
|
|
idev->last_fw_hb, lif->state[0]);
|
|
|
|
/* is this a transition? */
|
|
if (fw_status_ready != idev->fw_status_ready &&
|
|
!test_bit(IONIC_LIF_F_FW_STOPPING, lif->state)) {
|
|
bool trigger = false;
|
|
|
|
idev->fw_status_ready = fw_status_ready;
|
|
|
|
if (!fw_status_ready &&
|
|
!test_bit(IONIC_LIF_F_FW_RESET, lif->state) &&
|
|
!test_and_set_bit(IONIC_LIF_F_FW_STOPPING, lif->state)) {
|
|
dev_info(ionic->dev, "FW stopped 0x%02x\n", fw_status);
|
|
trigger = true;
|
|
|
|
} else if (fw_status_ready &&
|
|
test_bit(IONIC_LIF_F_FW_RESET, lif->state)) {
|
|
dev_info(ionic->dev, "FW running 0x%02x\n", fw_status);
|
|
trigger = true;
|
|
}
|
|
|
|
if (trigger) {
|
|
struct ionic_deferred_work *work;
|
|
|
|
work = kzalloc(sizeof(*work), GFP_ATOMIC);
|
|
if (work) {
|
|
work->type = IONIC_DW_TYPE_LIF_RESET;
|
|
work->fw_status = fw_status_ready;
|
|
ionic_lif_deferred_enqueue(&lif->deferred, work);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (!idev->fw_status_ready)
|
|
return -ENXIO;
|
|
|
|
/* Because of some variability in the actual FW heartbeat, we
|
|
* wait longer than the DEVCMD_TIMEOUT before checking again.
|
|
*/
|
|
last_check_time = idev->last_hb_time;
|
|
if (time_before(check_time, last_check_time + DEVCMD_TIMEOUT * 2 * HZ))
|
|
return 0;
|
|
|
|
fw_hb = ioread32(&idev->dev_info_regs->fw_heartbeat);
|
|
fw_hb_ready = fw_hb != idev->last_fw_hb;
|
|
|
|
/* early FW version had no heartbeat, so fake it */
|
|
if (!fw_hb_ready && !fw_hb)
|
|
fw_hb_ready = true;
|
|
|
|
dev_dbg(ionic->dev, "%s: fw_hb %u last_fw_hb %u ready %u\n",
|
|
__func__, fw_hb, idev->last_fw_hb, fw_hb_ready);
|
|
|
|
idev->last_fw_hb = fw_hb;
|
|
|
|
/* log a transition */
|
|
if (fw_hb_ready != idev->fw_hb_ready) {
|
|
idev->fw_hb_ready = fw_hb_ready;
|
|
if (!fw_hb_ready)
|
|
dev_info(ionic->dev, "FW heartbeat stalled at %d\n", fw_hb);
|
|
else
|
|
dev_info(ionic->dev, "FW heartbeat restored at %d\n", fw_hb);
|
|
}
|
|
|
|
if (!fw_hb_ready)
|
|
return -ENXIO;
|
|
|
|
idev->last_hb_time = check_time;
|
|
|
|
return 0;
|
|
}
|
|
|
|
u8 ionic_dev_cmd_status(struct ionic_dev *idev)
|
|
{
|
|
if (!idev->dev_cmd_regs)
|
|
return (u8)PCI_ERROR_RESPONSE;
|
|
return ioread8(&idev->dev_cmd_regs->comp.comp.status);
|
|
}
|
|
|
|
bool ionic_dev_cmd_done(struct ionic_dev *idev)
|
|
{
|
|
if (!idev->dev_cmd_regs)
|
|
return false;
|
|
return ioread32(&idev->dev_cmd_regs->done) & IONIC_DEV_CMD_DONE;
|
|
}
|
|
|
|
void ionic_dev_cmd_comp(struct ionic_dev *idev, union ionic_dev_cmd_comp *comp)
|
|
{
|
|
if (!idev->dev_cmd_regs)
|
|
return;
|
|
memcpy_fromio(comp, &idev->dev_cmd_regs->comp, sizeof(*comp));
|
|
}
|
|
|
|
void ionic_dev_cmd_go(struct ionic_dev *idev, union ionic_dev_cmd *cmd)
|
|
{
|
|
idev->opcode = cmd->cmd.opcode;
|
|
|
|
if (!idev->dev_cmd_regs)
|
|
return;
|
|
|
|
memcpy_toio(&idev->dev_cmd_regs->cmd, cmd, sizeof(*cmd));
|
|
iowrite32(0, &idev->dev_cmd_regs->done);
|
|
iowrite32(1, &idev->dev_cmd_regs->doorbell);
|
|
}
|
|
|
|
/* Device commands */
|
|
void ionic_dev_cmd_identify(struct ionic_dev *idev, u8 ver)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.identify.opcode = IONIC_CMD_IDENTIFY,
|
|
.identify.ver = ver,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_init(struct ionic_dev *idev)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.init.opcode = IONIC_CMD_INIT,
|
|
.init.type = 0,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_reset(struct ionic_dev *idev)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.reset.opcode = IONIC_CMD_RESET,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
/* Port commands */
|
|
void ionic_dev_cmd_port_identify(struct ionic_dev *idev)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.port_init.opcode = IONIC_CMD_PORT_IDENTIFY,
|
|
.port_init.index = 0,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_port_init(struct ionic_dev *idev)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.port_init.opcode = IONIC_CMD_PORT_INIT,
|
|
.port_init.index = 0,
|
|
.port_init.info_pa = cpu_to_le64(idev->port_info_pa),
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_port_reset(struct ionic_dev *idev)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.port_reset.opcode = IONIC_CMD_PORT_RESET,
|
|
.port_reset.index = 0,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_port_state(struct ionic_dev *idev, u8 state)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.port_setattr.opcode = IONIC_CMD_PORT_SETATTR,
|
|
.port_setattr.index = 0,
|
|
.port_setattr.attr = IONIC_PORT_ATTR_STATE,
|
|
.port_setattr.state = state,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_port_speed(struct ionic_dev *idev, u32 speed)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.port_setattr.opcode = IONIC_CMD_PORT_SETATTR,
|
|
.port_setattr.index = 0,
|
|
.port_setattr.attr = IONIC_PORT_ATTR_SPEED,
|
|
.port_setattr.speed = cpu_to_le32(speed),
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_port_autoneg(struct ionic_dev *idev, u8 an_enable)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.port_setattr.opcode = IONIC_CMD_PORT_SETATTR,
|
|
.port_setattr.index = 0,
|
|
.port_setattr.attr = IONIC_PORT_ATTR_AUTONEG,
|
|
.port_setattr.an_enable = an_enable,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_port_fec(struct ionic_dev *idev, u8 fec_type)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.port_setattr.opcode = IONIC_CMD_PORT_SETATTR,
|
|
.port_setattr.index = 0,
|
|
.port_setattr.attr = IONIC_PORT_ATTR_FEC,
|
|
.port_setattr.fec_type = fec_type,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_port_pause(struct ionic_dev *idev, u8 pause_type)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.port_setattr.opcode = IONIC_CMD_PORT_SETATTR,
|
|
.port_setattr.index = 0,
|
|
.port_setattr.attr = IONIC_PORT_ATTR_PAUSE,
|
|
.port_setattr.pause_type = pause_type,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
/* VF commands */
|
|
int ionic_set_vf_config(struct ionic *ionic, int vf,
|
|
struct ionic_vf_setattr_cmd *vfc)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.vf_setattr.opcode = IONIC_CMD_VF_SETATTR,
|
|
.vf_setattr.attr = vfc->attr,
|
|
.vf_setattr.vf_index = cpu_to_le16(vf),
|
|
};
|
|
int err;
|
|
|
|
memcpy(cmd.vf_setattr.pad, vfc->pad, sizeof(vfc->pad));
|
|
|
|
mutex_lock(&ionic->dev_cmd_lock);
|
|
ionic_dev_cmd_go(&ionic->idev, &cmd);
|
|
err = ionic_dev_cmd_wait(ionic, DEVCMD_TIMEOUT);
|
|
mutex_unlock(&ionic->dev_cmd_lock);
|
|
|
|
return err;
|
|
}
|
|
|
|
void ionic_vf_start(struct ionic *ionic)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.vf_ctrl.opcode = IONIC_CMD_VF_CTRL,
|
|
.vf_ctrl.ctrl_opcode = IONIC_VF_CTRL_START_ALL,
|
|
};
|
|
|
|
if (!(ionic->ident.dev.capabilities & cpu_to_le64(IONIC_DEV_CAP_VF_CTRL)))
|
|
return;
|
|
|
|
ionic_dev_cmd_go(&ionic->idev, &cmd);
|
|
ionic_dev_cmd_wait(ionic, DEVCMD_TIMEOUT);
|
|
}
|
|
|
|
/* LIF commands */
|
|
void ionic_dev_cmd_queue_identify(struct ionic_dev *idev,
|
|
u16 lif_type, u8 qtype, u8 qver)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.q_identify.opcode = IONIC_CMD_Q_IDENTIFY,
|
|
.q_identify.lif_type = cpu_to_le16(lif_type),
|
|
.q_identify.type = qtype,
|
|
.q_identify.ver = qver,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_lif_identify(struct ionic_dev *idev, u8 type, u8 ver)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.lif_identify.opcode = IONIC_CMD_LIF_IDENTIFY,
|
|
.lif_identify.type = type,
|
|
.lif_identify.ver = ver,
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_lif_init(struct ionic_dev *idev, u16 lif_index,
|
|
dma_addr_t info_pa)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.lif_init.opcode = IONIC_CMD_LIF_INIT,
|
|
.lif_init.index = cpu_to_le16(lif_index),
|
|
.lif_init.info_pa = cpu_to_le64(info_pa),
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_lif_reset(struct ionic_dev *idev, u16 lif_index)
|
|
{
|
|
union ionic_dev_cmd cmd = {
|
|
.lif_init.opcode = IONIC_CMD_LIF_RESET,
|
|
.lif_init.index = cpu_to_le16(lif_index),
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
void ionic_dev_cmd_adminq_init(struct ionic_dev *idev, struct ionic_qcq *qcq,
|
|
u16 lif_index, u16 intr_index)
|
|
{
|
|
struct ionic_queue *q = &qcq->q;
|
|
struct ionic_cq *cq = &qcq->cq;
|
|
|
|
union ionic_dev_cmd cmd = {
|
|
.q_init.opcode = IONIC_CMD_Q_INIT,
|
|
.q_init.lif_index = cpu_to_le16(lif_index),
|
|
.q_init.type = q->type,
|
|
.q_init.ver = qcq->q.lif->qtype_info[q->type].version,
|
|
.q_init.index = cpu_to_le32(q->index),
|
|
.q_init.flags = cpu_to_le16(IONIC_QINIT_F_IRQ |
|
|
IONIC_QINIT_F_ENA),
|
|
.q_init.pid = cpu_to_le16(q->pid),
|
|
.q_init.intr_index = cpu_to_le16(intr_index),
|
|
.q_init.ring_size = ilog2(q->num_descs),
|
|
.q_init.ring_base = cpu_to_le64(q->base_pa),
|
|
.q_init.cq_ring_base = cpu_to_le64(cq->base_pa),
|
|
};
|
|
|
|
ionic_dev_cmd_go(idev, &cmd);
|
|
}
|
|
|
|
int ionic_db_page_num(struct ionic_lif *lif, int pid)
|
|
{
|
|
return (lif->hw_index * lif->dbid_count) + pid;
|
|
}
|
|
|
|
int ionic_get_cmb(struct ionic_lif *lif, u32 *pgid, phys_addr_t *pgaddr, int order)
|
|
{
|
|
struct ionic_dev *idev = &lif->ionic->idev;
|
|
int ret;
|
|
|
|
mutex_lock(&idev->cmb_inuse_lock);
|
|
ret = bitmap_find_free_region(idev->cmb_inuse, idev->cmb_npages, order);
|
|
mutex_unlock(&idev->cmb_inuse_lock);
|
|
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
*pgid = ret;
|
|
*pgaddr = idev->phy_cmb_pages + ret * PAGE_SIZE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ionic_put_cmb(struct ionic_lif *lif, u32 pgid, int order)
|
|
{
|
|
struct ionic_dev *idev = &lif->ionic->idev;
|
|
|
|
mutex_lock(&idev->cmb_inuse_lock);
|
|
bitmap_release_region(idev->cmb_inuse, pgid, order);
|
|
mutex_unlock(&idev->cmb_inuse_lock);
|
|
}
|
|
|
|
int ionic_cq_init(struct ionic_lif *lif, struct ionic_cq *cq,
|
|
struct ionic_intr_info *intr,
|
|
unsigned int num_descs, size_t desc_size)
|
|
{
|
|
unsigned int ring_size;
|
|
|
|
if (desc_size == 0 || !is_power_of_2(num_descs))
|
|
return -EINVAL;
|
|
|
|
ring_size = ilog2(num_descs);
|
|
if (ring_size < 2 || ring_size > 16)
|
|
return -EINVAL;
|
|
|
|
cq->lif = lif;
|
|
cq->bound_intr = intr;
|
|
cq->num_descs = num_descs;
|
|
cq->desc_size = desc_size;
|
|
cq->tail_idx = 0;
|
|
cq->done_color = 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ionic_cq_map(struct ionic_cq *cq, void *base, dma_addr_t base_pa)
|
|
{
|
|
struct ionic_cq_info *cur;
|
|
unsigned int i;
|
|
|
|
cq->base = base;
|
|
cq->base_pa = base_pa;
|
|
|
|
for (i = 0, cur = cq->info; i < cq->num_descs; i++, cur++)
|
|
cur->cq_desc = base + (i * cq->desc_size);
|
|
}
|
|
|
|
void ionic_cq_bind(struct ionic_cq *cq, struct ionic_queue *q)
|
|
{
|
|
cq->bound_q = q;
|
|
}
|
|
|
|
unsigned int ionic_cq_service(struct ionic_cq *cq, unsigned int work_to_do,
|
|
ionic_cq_cb cb, ionic_cq_done_cb done_cb,
|
|
void *done_arg)
|
|
{
|
|
struct ionic_cq_info *cq_info;
|
|
unsigned int work_done = 0;
|
|
|
|
if (work_to_do == 0)
|
|
return 0;
|
|
|
|
cq_info = &cq->info[cq->tail_idx];
|
|
while (cb(cq, cq_info)) {
|
|
if (cq->tail_idx == cq->num_descs - 1)
|
|
cq->done_color = !cq->done_color;
|
|
cq->tail_idx = (cq->tail_idx + 1) & (cq->num_descs - 1);
|
|
cq_info = &cq->info[cq->tail_idx];
|
|
|
|
if (++work_done >= work_to_do)
|
|
break;
|
|
}
|
|
|
|
if (work_done && done_cb)
|
|
done_cb(done_arg);
|
|
|
|
return work_done;
|
|
}
|
|
|
|
int ionic_q_init(struct ionic_lif *lif, struct ionic_dev *idev,
|
|
struct ionic_queue *q, unsigned int index, const char *name,
|
|
unsigned int num_descs, size_t desc_size,
|
|
size_t sg_desc_size, unsigned int pid)
|
|
{
|
|
unsigned int ring_size;
|
|
|
|
if (desc_size == 0 || !is_power_of_2(num_descs))
|
|
return -EINVAL;
|
|
|
|
ring_size = ilog2(num_descs);
|
|
if (ring_size < 2 || ring_size > 16)
|
|
return -EINVAL;
|
|
|
|
q->lif = lif;
|
|
q->idev = idev;
|
|
q->index = index;
|
|
q->num_descs = num_descs;
|
|
q->desc_size = desc_size;
|
|
q->sg_desc_size = sg_desc_size;
|
|
q->tail_idx = 0;
|
|
q->head_idx = 0;
|
|
q->pid = pid;
|
|
|
|
snprintf(q->name, sizeof(q->name), "L%d-%s%u", lif->index, name, index);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ionic_q_map(struct ionic_queue *q, void *base, dma_addr_t base_pa)
|
|
{
|
|
struct ionic_desc_info *cur;
|
|
unsigned int i;
|
|
|
|
q->base = base;
|
|
q->base_pa = base_pa;
|
|
|
|
for (i = 0, cur = q->info; i < q->num_descs; i++, cur++)
|
|
cur->desc = base + (i * q->desc_size);
|
|
}
|
|
|
|
void ionic_q_cmb_map(struct ionic_queue *q, void __iomem *base, dma_addr_t base_pa)
|
|
{
|
|
struct ionic_desc_info *cur;
|
|
unsigned int i;
|
|
|
|
q->cmb_base = base;
|
|
q->cmb_base_pa = base_pa;
|
|
|
|
for (i = 0, cur = q->info; i < q->num_descs; i++, cur++)
|
|
cur->cmb_desc = base + (i * q->desc_size);
|
|
}
|
|
|
|
void ionic_q_sg_map(struct ionic_queue *q, void *base, dma_addr_t base_pa)
|
|
{
|
|
struct ionic_desc_info *cur;
|
|
unsigned int i;
|
|
|
|
q->sg_base = base;
|
|
q->sg_base_pa = base_pa;
|
|
|
|
for (i = 0, cur = q->info; i < q->num_descs; i++, cur++)
|
|
cur->sg_desc = base + (i * q->sg_desc_size);
|
|
}
|
|
|
|
void ionic_q_post(struct ionic_queue *q, bool ring_doorbell, ionic_desc_cb cb,
|
|
void *cb_arg)
|
|
{
|
|
struct ionic_desc_info *desc_info;
|
|
struct ionic_lif *lif = q->lif;
|
|
struct device *dev = q->dev;
|
|
|
|
desc_info = &q->info[q->head_idx];
|
|
desc_info->cb = cb;
|
|
desc_info->cb_arg = cb_arg;
|
|
|
|
q->head_idx = (q->head_idx + 1) & (q->num_descs - 1);
|
|
|
|
dev_dbg(dev, "lif=%d qname=%s qid=%d qtype=%d p_index=%d ringdb=%d\n",
|
|
q->lif->index, q->name, q->hw_type, q->hw_index,
|
|
q->head_idx, ring_doorbell);
|
|
|
|
if (ring_doorbell) {
|
|
ionic_dbell_ring(lif->kern_dbpage, q->hw_type,
|
|
q->dbval | q->head_idx);
|
|
|
|
q->dbell_jiffies = jiffies;
|
|
|
|
if (q_to_qcq(q)->napi_qcq)
|
|
mod_timer(&q_to_qcq(q)->napi_qcq->napi_deadline,
|
|
jiffies + IONIC_NAPI_DEADLINE);
|
|
}
|
|
}
|
|
|
|
static bool ionic_q_is_posted(struct ionic_queue *q, unsigned int pos)
|
|
{
|
|
unsigned int mask, tail, head;
|
|
|
|
mask = q->num_descs - 1;
|
|
tail = q->tail_idx;
|
|
head = q->head_idx;
|
|
|
|
return ((pos - tail) & mask) < ((head - tail) & mask);
|
|
}
|
|
|
|
void ionic_q_service(struct ionic_queue *q, struct ionic_cq_info *cq_info,
|
|
unsigned int stop_index)
|
|
{
|
|
struct ionic_desc_info *desc_info;
|
|
ionic_desc_cb cb;
|
|
void *cb_arg;
|
|
u16 index;
|
|
|
|
/* check for empty queue */
|
|
if (q->tail_idx == q->head_idx)
|
|
return;
|
|
|
|
/* stop index must be for a descriptor that is not yet completed */
|
|
if (unlikely(!ionic_q_is_posted(q, stop_index)))
|
|
dev_err(q->dev,
|
|
"ionic stop is not posted %s stop %u tail %u head %u\n",
|
|
q->name, stop_index, q->tail_idx, q->head_idx);
|
|
|
|
do {
|
|
desc_info = &q->info[q->tail_idx];
|
|
index = q->tail_idx;
|
|
q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1);
|
|
|
|
cb = desc_info->cb;
|
|
cb_arg = desc_info->cb_arg;
|
|
|
|
desc_info->cb = NULL;
|
|
desc_info->cb_arg = NULL;
|
|
|
|
if (cb)
|
|
cb(q, desc_info, cq_info, cb_arg);
|
|
} while (index != stop_index);
|
|
}
|