1375 lines
41 KiB
C
1375 lines
41 KiB
C
// SPDX-License-Identifier: BSD-3-Clause-Clear
|
|
/*
|
|
* Copyright (c) 2018-2019 The Linux Foundation. All rights reserved.
|
|
*/
|
|
#include <linux/dma-mapping.h>
|
|
#include "hal_tx.h"
|
|
#include "debug.h"
|
|
#include "hal_desc.h"
|
|
#include "hif.h"
|
|
|
|
static const struct hal_srng_config hw_srng_config_template[] = {
|
|
/* TODO: max_rings can populated by querying HW capabilities */
|
|
{ /* REO_DST */
|
|
.start_ring_id = HAL_SRNG_RING_ID_REO2SW1,
|
|
.max_rings = 4,
|
|
.entry_size = sizeof(struct hal_reo_dest_ring) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_DST,
|
|
.max_size = HAL_REO_REO2SW1_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* REO_EXCEPTION */
|
|
/* Designating REO2TCL ring as exception ring. This ring is
|
|
* similar to other REO2SW rings though it is named as REO2TCL.
|
|
* Any of theREO2SW rings can be used as exception ring.
|
|
*/
|
|
.start_ring_id = HAL_SRNG_RING_ID_REO2TCL,
|
|
.max_rings = 1,
|
|
.entry_size = sizeof(struct hal_reo_dest_ring) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_DST,
|
|
.max_size = HAL_REO_REO2TCL_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* REO_REINJECT */
|
|
.start_ring_id = HAL_SRNG_RING_ID_SW2REO,
|
|
.max_rings = 1,
|
|
.entry_size = sizeof(struct hal_reo_entrance_ring) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_REO_SW2REO_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* REO_CMD */
|
|
.start_ring_id = HAL_SRNG_RING_ID_REO_CMD,
|
|
.max_rings = 1,
|
|
.entry_size = (sizeof(struct hal_tlv_hdr) +
|
|
sizeof(struct hal_reo_get_queue_stats)) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_REO_CMD_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* REO_STATUS */
|
|
.start_ring_id = HAL_SRNG_RING_ID_REO_STATUS,
|
|
.max_rings = 1,
|
|
.entry_size = (sizeof(struct hal_tlv_hdr) +
|
|
sizeof(struct hal_reo_get_queue_stats_status)) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_DST,
|
|
.max_size = HAL_REO_STATUS_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* TCL_DATA */
|
|
.start_ring_id = HAL_SRNG_RING_ID_SW2TCL1,
|
|
.max_rings = 3,
|
|
.entry_size = (sizeof(struct hal_tlv_hdr) +
|
|
sizeof(struct hal_tcl_data_cmd)) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_SW2TCL1_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* TCL_CMD */
|
|
.start_ring_id = HAL_SRNG_RING_ID_SW2TCL_CMD,
|
|
.max_rings = 1,
|
|
.entry_size = (sizeof(struct hal_tlv_hdr) +
|
|
sizeof(struct hal_tcl_gse_cmd)) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_SW2TCL1_CMD_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* TCL_STATUS */
|
|
.start_ring_id = HAL_SRNG_RING_ID_TCL_STATUS,
|
|
.max_rings = 1,
|
|
.entry_size = (sizeof(struct hal_tlv_hdr) +
|
|
sizeof(struct hal_tcl_status_ring)) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_DST,
|
|
.max_size = HAL_TCL_STATUS_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* CE_SRC */
|
|
.start_ring_id = HAL_SRNG_RING_ID_CE0_SRC,
|
|
.max_rings = 12,
|
|
.entry_size = sizeof(struct hal_ce_srng_src_desc) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_CE_SRC_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* CE_DST */
|
|
.start_ring_id = HAL_SRNG_RING_ID_CE0_DST,
|
|
.max_rings = 12,
|
|
.entry_size = sizeof(struct hal_ce_srng_dest_desc) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_CE_DST_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* CE_DST_STATUS */
|
|
.start_ring_id = HAL_SRNG_RING_ID_CE0_DST_STATUS,
|
|
.max_rings = 12,
|
|
.entry_size = sizeof(struct hal_ce_srng_dst_status_desc) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_DST,
|
|
.max_size = HAL_CE_DST_STATUS_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* WBM_IDLE_LINK */
|
|
.start_ring_id = HAL_SRNG_RING_ID_WBM_IDLE_LINK,
|
|
.max_rings = 1,
|
|
.entry_size = sizeof(struct hal_wbm_link_desc) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_WBM_IDLE_LINK_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* SW2WBM_RELEASE */
|
|
.start_ring_id = HAL_SRNG_RING_ID_WBM_SW_RELEASE,
|
|
.max_rings = 1,
|
|
.entry_size = sizeof(struct hal_wbm_release_ring) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_SW2WBM_RELEASE_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* WBM2SW_RELEASE */
|
|
.start_ring_id = HAL_SRNG_RING_ID_WBM2SW0_RELEASE,
|
|
.max_rings = 4,
|
|
.entry_size = sizeof(struct hal_wbm_release_ring) >> 2,
|
|
.lmac_ring = false,
|
|
.ring_dir = HAL_SRNG_DIR_DST,
|
|
.max_size = HAL_WBM2SW_RELEASE_RING_BASE_MSB_RING_SIZE,
|
|
},
|
|
{ /* RXDMA_BUF */
|
|
.start_ring_id = HAL_SRNG_RING_ID_WMAC1_SW2RXDMA0_BUF,
|
|
.max_rings = 2,
|
|
.entry_size = sizeof(struct hal_wbm_buffer_ring) >> 2,
|
|
.lmac_ring = true,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_RXDMA_RING_MAX_SIZE,
|
|
},
|
|
{ /* RXDMA_DST */
|
|
.start_ring_id = HAL_SRNG_RING_ID_WMAC1_RXDMA2SW0,
|
|
.max_rings = 1,
|
|
.entry_size = sizeof(struct hal_reo_entrance_ring) >> 2,
|
|
.lmac_ring = true,
|
|
.ring_dir = HAL_SRNG_DIR_DST,
|
|
.max_size = HAL_RXDMA_RING_MAX_SIZE,
|
|
},
|
|
{ /* RXDMA_MONITOR_BUF */
|
|
.start_ring_id = HAL_SRNG_RING_ID_WMAC1_SW2RXDMA2_BUF,
|
|
.max_rings = 1,
|
|
.entry_size = sizeof(struct hal_wbm_buffer_ring) >> 2,
|
|
.lmac_ring = true,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_RXDMA_RING_MAX_SIZE,
|
|
},
|
|
{ /* RXDMA_MONITOR_STATUS */
|
|
.start_ring_id = HAL_SRNG_RING_ID_WMAC1_SW2RXDMA1_STATBUF,
|
|
.max_rings = 1,
|
|
.entry_size = sizeof(struct hal_wbm_buffer_ring) >> 2,
|
|
.lmac_ring = true,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_RXDMA_RING_MAX_SIZE,
|
|
},
|
|
{ /* RXDMA_MONITOR_DST */
|
|
.start_ring_id = HAL_SRNG_RING_ID_WMAC1_RXDMA2SW1,
|
|
.max_rings = 1,
|
|
.entry_size = sizeof(struct hal_reo_entrance_ring) >> 2,
|
|
.lmac_ring = true,
|
|
.ring_dir = HAL_SRNG_DIR_DST,
|
|
.max_size = HAL_RXDMA_RING_MAX_SIZE,
|
|
},
|
|
{ /* RXDMA_MONITOR_DESC */
|
|
.start_ring_id = HAL_SRNG_RING_ID_WMAC1_SW2RXDMA1_DESC,
|
|
.max_rings = 1,
|
|
.entry_size = sizeof(struct hal_wbm_buffer_ring) >> 2,
|
|
.lmac_ring = true,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_RXDMA_RING_MAX_SIZE,
|
|
},
|
|
{ /* RXDMA DIR BUF */
|
|
.start_ring_id = HAL_SRNG_RING_ID_RXDMA_DIR_BUF,
|
|
.max_rings = 1,
|
|
.entry_size = 8 >> 2, /* TODO: Define the struct */
|
|
.lmac_ring = true,
|
|
.ring_dir = HAL_SRNG_DIR_SRC,
|
|
.max_size = HAL_RXDMA_RING_MAX_SIZE,
|
|
},
|
|
};
|
|
|
|
static int ath11k_hal_alloc_cont_rdp(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
size_t size;
|
|
|
|
size = sizeof(u32) * HAL_SRNG_RING_ID_MAX;
|
|
hal->rdp.vaddr = dma_alloc_coherent(ab->dev, size, &hal->rdp.paddr,
|
|
GFP_KERNEL);
|
|
if (!hal->rdp.vaddr)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath11k_hal_free_cont_rdp(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
size_t size;
|
|
|
|
if (!hal->rdp.vaddr)
|
|
return;
|
|
|
|
size = sizeof(u32) * HAL_SRNG_RING_ID_MAX;
|
|
dma_free_coherent(ab->dev, size,
|
|
hal->rdp.vaddr, hal->rdp.paddr);
|
|
hal->rdp.vaddr = NULL;
|
|
}
|
|
|
|
static int ath11k_hal_alloc_cont_wrp(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
size_t size;
|
|
|
|
size = sizeof(u32) * HAL_SRNG_NUM_LMAC_RINGS;
|
|
hal->wrp.vaddr = dma_alloc_coherent(ab->dev, size, &hal->wrp.paddr,
|
|
GFP_KERNEL);
|
|
if (!hal->wrp.vaddr)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath11k_hal_free_cont_wrp(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
size_t size;
|
|
|
|
if (!hal->wrp.vaddr)
|
|
return;
|
|
|
|
size = sizeof(u32) * HAL_SRNG_NUM_LMAC_RINGS;
|
|
dma_free_coherent(ab->dev, size,
|
|
hal->wrp.vaddr, hal->wrp.paddr);
|
|
hal->wrp.vaddr = NULL;
|
|
}
|
|
|
|
static void ath11k_hal_ce_dst_setup(struct ath11k_base *ab,
|
|
struct hal_srng *srng, int ring_num)
|
|
{
|
|
struct hal_srng_config *srng_config = &ab->hal.srng_config[HAL_CE_DST];
|
|
u32 addr;
|
|
u32 val;
|
|
|
|
addr = HAL_CE_DST_RING_CTRL +
|
|
srng_config->reg_start[HAL_SRNG_REG_GRP_R0] +
|
|
ring_num * srng_config->reg_size[HAL_SRNG_REG_GRP_R0];
|
|
|
|
val = ath11k_hif_read32(ab, addr);
|
|
val &= ~HAL_CE_DST_R0_DEST_CTRL_MAX_LEN;
|
|
val |= FIELD_PREP(HAL_CE_DST_R0_DEST_CTRL_MAX_LEN,
|
|
srng->u.dst_ring.max_buffer_length);
|
|
ath11k_hif_write32(ab, addr, val);
|
|
}
|
|
|
|
static void ath11k_hal_srng_dst_hw_init(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
u32 val;
|
|
u64 hp_addr;
|
|
u32 reg_base;
|
|
|
|
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R0];
|
|
|
|
if (srng->flags & HAL_SRNG_FLAGS_MSI_INTR) {
|
|
ath11k_hif_write32(ab, reg_base +
|
|
HAL_REO1_RING_MSI1_BASE_LSB_OFFSET(ab),
|
|
srng->msi_addr);
|
|
|
|
val = FIELD_PREP(HAL_REO1_RING_MSI1_BASE_MSB_ADDR,
|
|
((u64)srng->msi_addr >>
|
|
HAL_ADDR_MSB_REG_SHIFT)) |
|
|
HAL_REO1_RING_MSI1_BASE_MSB_MSI1_ENABLE;
|
|
ath11k_hif_write32(ab, reg_base +
|
|
HAL_REO1_RING_MSI1_BASE_MSB_OFFSET(ab), val);
|
|
|
|
ath11k_hif_write32(ab,
|
|
reg_base + HAL_REO1_RING_MSI1_DATA_OFFSET(ab),
|
|
srng->msi_data);
|
|
}
|
|
|
|
ath11k_hif_write32(ab, reg_base, srng->ring_base_paddr);
|
|
|
|
val = FIELD_PREP(HAL_REO1_RING_BASE_MSB_RING_BASE_ADDR_MSB,
|
|
((u64)srng->ring_base_paddr >>
|
|
HAL_ADDR_MSB_REG_SHIFT)) |
|
|
FIELD_PREP(HAL_REO1_RING_BASE_MSB_RING_SIZE,
|
|
(srng->entry_size * srng->num_entries));
|
|
ath11k_hif_write32(ab, reg_base + HAL_REO1_RING_BASE_MSB_OFFSET(ab), val);
|
|
|
|
val = FIELD_PREP(HAL_REO1_RING_ID_RING_ID, srng->ring_id) |
|
|
FIELD_PREP(HAL_REO1_RING_ID_ENTRY_SIZE, srng->entry_size);
|
|
ath11k_hif_write32(ab, reg_base + HAL_REO1_RING_ID_OFFSET(ab), val);
|
|
|
|
/* interrupt setup */
|
|
val = FIELD_PREP(HAL_REO1_RING_PRDR_INT_SETUP_INTR_TMR_THOLD,
|
|
(srng->intr_timer_thres_us >> 3));
|
|
|
|
val |= FIELD_PREP(HAL_REO1_RING_PRDR_INT_SETUP_BATCH_COUNTER_THOLD,
|
|
(srng->intr_batch_cntr_thres_entries *
|
|
srng->entry_size));
|
|
|
|
ath11k_hif_write32(ab,
|
|
reg_base + HAL_REO1_RING_PRODUCER_INT_SETUP_OFFSET(ab),
|
|
val);
|
|
|
|
hp_addr = hal->rdp.paddr +
|
|
((unsigned long)srng->u.dst_ring.hp_addr -
|
|
(unsigned long)hal->rdp.vaddr);
|
|
ath11k_hif_write32(ab, reg_base + HAL_REO1_RING_HP_ADDR_LSB_OFFSET(ab),
|
|
hp_addr & HAL_ADDR_LSB_REG_MASK);
|
|
ath11k_hif_write32(ab, reg_base + HAL_REO1_RING_HP_ADDR_MSB_OFFSET(ab),
|
|
hp_addr >> HAL_ADDR_MSB_REG_SHIFT);
|
|
|
|
/* Initialize head and tail pointers to indicate ring is empty */
|
|
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R2];
|
|
ath11k_hif_write32(ab, reg_base, 0);
|
|
ath11k_hif_write32(ab, reg_base + HAL_REO1_RING_TP_OFFSET(ab), 0);
|
|
*srng->u.dst_ring.hp_addr = 0;
|
|
|
|
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R0];
|
|
val = 0;
|
|
if (srng->flags & HAL_SRNG_FLAGS_DATA_TLV_SWAP)
|
|
val |= HAL_REO1_RING_MISC_DATA_TLV_SWAP;
|
|
if (srng->flags & HAL_SRNG_FLAGS_RING_PTR_SWAP)
|
|
val |= HAL_REO1_RING_MISC_HOST_FW_SWAP;
|
|
if (srng->flags & HAL_SRNG_FLAGS_MSI_SWAP)
|
|
val |= HAL_REO1_RING_MISC_MSI_SWAP;
|
|
val |= HAL_REO1_RING_MISC_SRNG_ENABLE;
|
|
|
|
ath11k_hif_write32(ab, reg_base + HAL_REO1_RING_MISC_OFFSET(ab), val);
|
|
}
|
|
|
|
static void ath11k_hal_srng_src_hw_init(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
u32 val;
|
|
u64 tp_addr;
|
|
u32 reg_base;
|
|
|
|
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R0];
|
|
|
|
if (srng->flags & HAL_SRNG_FLAGS_MSI_INTR) {
|
|
ath11k_hif_write32(ab, reg_base +
|
|
HAL_TCL1_RING_MSI1_BASE_LSB_OFFSET(ab),
|
|
srng->msi_addr);
|
|
|
|
val = FIELD_PREP(HAL_TCL1_RING_MSI1_BASE_MSB_ADDR,
|
|
((u64)srng->msi_addr >>
|
|
HAL_ADDR_MSB_REG_SHIFT)) |
|
|
HAL_TCL1_RING_MSI1_BASE_MSB_MSI1_ENABLE;
|
|
ath11k_hif_write32(ab, reg_base +
|
|
HAL_TCL1_RING_MSI1_BASE_MSB_OFFSET(ab),
|
|
val);
|
|
|
|
ath11k_hif_write32(ab, reg_base +
|
|
HAL_TCL1_RING_MSI1_DATA_OFFSET(ab),
|
|
srng->msi_data);
|
|
}
|
|
|
|
ath11k_hif_write32(ab, reg_base, srng->ring_base_paddr);
|
|
|
|
val = FIELD_PREP(HAL_TCL1_RING_BASE_MSB_RING_BASE_ADDR_MSB,
|
|
((u64)srng->ring_base_paddr >>
|
|
HAL_ADDR_MSB_REG_SHIFT)) |
|
|
FIELD_PREP(HAL_TCL1_RING_BASE_MSB_RING_SIZE,
|
|
(srng->entry_size * srng->num_entries));
|
|
ath11k_hif_write32(ab, reg_base + HAL_TCL1_RING_BASE_MSB_OFFSET(ab), val);
|
|
|
|
val = FIELD_PREP(HAL_REO1_RING_ID_ENTRY_SIZE, srng->entry_size);
|
|
ath11k_hif_write32(ab, reg_base + HAL_TCL1_RING_ID_OFFSET(ab), val);
|
|
|
|
if (srng->ring_id == HAL_SRNG_RING_ID_WBM_IDLE_LINK) {
|
|
ath11k_hif_write32(ab, reg_base, (u32)srng->ring_base_paddr);
|
|
val = FIELD_PREP(HAL_TCL1_RING_BASE_MSB_RING_BASE_ADDR_MSB,
|
|
((u64)srng->ring_base_paddr >>
|
|
HAL_ADDR_MSB_REG_SHIFT)) |
|
|
FIELD_PREP(HAL_TCL1_RING_BASE_MSB_RING_SIZE,
|
|
(srng->entry_size * srng->num_entries));
|
|
ath11k_hif_write32(ab, reg_base + HAL_TCL1_RING_BASE_MSB_OFFSET(ab), val);
|
|
}
|
|
|
|
/* interrupt setup */
|
|
/* NOTE: IPQ8074 v2 requires the interrupt timer threshold in the
|
|
* unit of 8 usecs instead of 1 usec (as required by v1).
|
|
*/
|
|
val = FIELD_PREP(HAL_TCL1_RING_CONSR_INT_SETUP_IX0_INTR_TMR_THOLD,
|
|
srng->intr_timer_thres_us);
|
|
|
|
val |= FIELD_PREP(HAL_TCL1_RING_CONSR_INT_SETUP_IX0_BATCH_COUNTER_THOLD,
|
|
(srng->intr_batch_cntr_thres_entries *
|
|
srng->entry_size));
|
|
|
|
ath11k_hif_write32(ab,
|
|
reg_base + HAL_TCL1_RING_CONSR_INT_SETUP_IX0_OFFSET(ab),
|
|
val);
|
|
|
|
val = 0;
|
|
if (srng->flags & HAL_SRNG_FLAGS_LOW_THRESH_INTR_EN) {
|
|
val |= FIELD_PREP(HAL_TCL1_RING_CONSR_INT_SETUP_IX1_LOW_THOLD,
|
|
srng->u.src_ring.low_threshold);
|
|
}
|
|
ath11k_hif_write32(ab,
|
|
reg_base + HAL_TCL1_RING_CONSR_INT_SETUP_IX1_OFFSET(ab),
|
|
val);
|
|
|
|
if (srng->ring_id != HAL_SRNG_RING_ID_WBM_IDLE_LINK) {
|
|
tp_addr = hal->rdp.paddr +
|
|
((unsigned long)srng->u.src_ring.tp_addr -
|
|
(unsigned long)hal->rdp.vaddr);
|
|
ath11k_hif_write32(ab,
|
|
reg_base + HAL_TCL1_RING_TP_ADDR_LSB_OFFSET(ab),
|
|
tp_addr & HAL_ADDR_LSB_REG_MASK);
|
|
ath11k_hif_write32(ab,
|
|
reg_base + HAL_TCL1_RING_TP_ADDR_MSB_OFFSET(ab),
|
|
tp_addr >> HAL_ADDR_MSB_REG_SHIFT);
|
|
}
|
|
|
|
/* Initialize head and tail pointers to indicate ring is empty */
|
|
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R2];
|
|
ath11k_hif_write32(ab, reg_base, 0);
|
|
ath11k_hif_write32(ab, reg_base + HAL_TCL1_RING_TP_OFFSET, 0);
|
|
*srng->u.src_ring.tp_addr = 0;
|
|
|
|
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R0];
|
|
val = 0;
|
|
if (srng->flags & HAL_SRNG_FLAGS_DATA_TLV_SWAP)
|
|
val |= HAL_TCL1_RING_MISC_DATA_TLV_SWAP;
|
|
if (srng->flags & HAL_SRNG_FLAGS_RING_PTR_SWAP)
|
|
val |= HAL_TCL1_RING_MISC_HOST_FW_SWAP;
|
|
if (srng->flags & HAL_SRNG_FLAGS_MSI_SWAP)
|
|
val |= HAL_TCL1_RING_MISC_MSI_SWAP;
|
|
|
|
/* Loop count is not used for SRC rings */
|
|
val |= HAL_TCL1_RING_MISC_MSI_LOOPCNT_DISABLE;
|
|
|
|
val |= HAL_TCL1_RING_MISC_SRNG_ENABLE;
|
|
|
|
ath11k_hif_write32(ab, reg_base + HAL_TCL1_RING_MISC_OFFSET(ab), val);
|
|
}
|
|
|
|
static void ath11k_hal_srng_hw_init(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
if (srng->ring_dir == HAL_SRNG_DIR_SRC)
|
|
ath11k_hal_srng_src_hw_init(ab, srng);
|
|
else
|
|
ath11k_hal_srng_dst_hw_init(ab, srng);
|
|
}
|
|
|
|
static int ath11k_hal_srng_get_ring_id(struct ath11k_base *ab,
|
|
enum hal_ring_type type,
|
|
int ring_num, int mac_id)
|
|
{
|
|
struct hal_srng_config *srng_config = &ab->hal.srng_config[type];
|
|
int ring_id;
|
|
|
|
if (ring_num >= srng_config->max_rings) {
|
|
ath11k_warn(ab, "invalid ring number :%d\n", ring_num);
|
|
return -EINVAL;
|
|
}
|
|
|
|
ring_id = srng_config->start_ring_id + ring_num;
|
|
if (srng_config->lmac_ring)
|
|
ring_id += mac_id * HAL_SRNG_RINGS_PER_LMAC;
|
|
|
|
if (WARN_ON(ring_id >= HAL_SRNG_RING_ID_MAX))
|
|
return -EINVAL;
|
|
|
|
return ring_id;
|
|
}
|
|
|
|
int ath11k_hal_srng_get_entrysize(struct ath11k_base *ab, u32 ring_type)
|
|
{
|
|
struct hal_srng_config *srng_config;
|
|
|
|
if (WARN_ON(ring_type >= HAL_MAX_RING_TYPES))
|
|
return -EINVAL;
|
|
|
|
srng_config = &ab->hal.srng_config[ring_type];
|
|
|
|
return (srng_config->entry_size << 2);
|
|
}
|
|
|
|
int ath11k_hal_srng_get_max_entries(struct ath11k_base *ab, u32 ring_type)
|
|
{
|
|
struct hal_srng_config *srng_config;
|
|
|
|
if (WARN_ON(ring_type >= HAL_MAX_RING_TYPES))
|
|
return -EINVAL;
|
|
|
|
srng_config = &ab->hal.srng_config[ring_type];
|
|
|
|
return (srng_config->max_size / srng_config->entry_size);
|
|
}
|
|
|
|
void ath11k_hal_srng_get_params(struct ath11k_base *ab, struct hal_srng *srng,
|
|
struct hal_srng_params *params)
|
|
{
|
|
params->ring_base_paddr = srng->ring_base_paddr;
|
|
params->ring_base_vaddr = srng->ring_base_vaddr;
|
|
params->num_entries = srng->num_entries;
|
|
params->intr_timer_thres_us = srng->intr_timer_thres_us;
|
|
params->intr_batch_cntr_thres_entries =
|
|
srng->intr_batch_cntr_thres_entries;
|
|
params->low_threshold = srng->u.src_ring.low_threshold;
|
|
params->msi_addr = srng->msi_addr;
|
|
params->msi_data = srng->msi_data;
|
|
params->flags = srng->flags;
|
|
}
|
|
|
|
dma_addr_t ath11k_hal_srng_get_hp_addr(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
if (!(srng->flags & HAL_SRNG_FLAGS_LMAC_RING))
|
|
return 0;
|
|
|
|
if (srng->ring_dir == HAL_SRNG_DIR_SRC)
|
|
return ab->hal.wrp.paddr +
|
|
((unsigned long)srng->u.src_ring.hp_addr -
|
|
(unsigned long)ab->hal.wrp.vaddr);
|
|
else
|
|
return ab->hal.rdp.paddr +
|
|
((unsigned long)srng->u.dst_ring.hp_addr -
|
|
(unsigned long)ab->hal.rdp.vaddr);
|
|
}
|
|
|
|
dma_addr_t ath11k_hal_srng_get_tp_addr(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
if (!(srng->flags & HAL_SRNG_FLAGS_LMAC_RING))
|
|
return 0;
|
|
|
|
if (srng->ring_dir == HAL_SRNG_DIR_SRC)
|
|
return ab->hal.rdp.paddr +
|
|
((unsigned long)srng->u.src_ring.tp_addr -
|
|
(unsigned long)ab->hal.rdp.vaddr);
|
|
else
|
|
return ab->hal.wrp.paddr +
|
|
((unsigned long)srng->u.dst_ring.tp_addr -
|
|
(unsigned long)ab->hal.wrp.vaddr);
|
|
}
|
|
|
|
u32 ath11k_hal_ce_get_desc_size(enum hal_ce_desc type)
|
|
{
|
|
switch (type) {
|
|
case HAL_CE_DESC_SRC:
|
|
return sizeof(struct hal_ce_srng_src_desc);
|
|
case HAL_CE_DESC_DST:
|
|
return sizeof(struct hal_ce_srng_dest_desc);
|
|
case HAL_CE_DESC_DST_STATUS:
|
|
return sizeof(struct hal_ce_srng_dst_status_desc);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ath11k_hal_ce_src_set_desc(void *buf, dma_addr_t paddr, u32 len, u32 id,
|
|
u8 byte_swap_data)
|
|
{
|
|
struct hal_ce_srng_src_desc *desc = (struct hal_ce_srng_src_desc *)buf;
|
|
|
|
desc->buffer_addr_low = paddr & HAL_ADDR_LSB_REG_MASK;
|
|
desc->buffer_addr_info =
|
|
FIELD_PREP(HAL_CE_SRC_DESC_ADDR_INFO_ADDR_HI,
|
|
((u64)paddr >> HAL_ADDR_MSB_REG_SHIFT)) |
|
|
FIELD_PREP(HAL_CE_SRC_DESC_ADDR_INFO_BYTE_SWAP,
|
|
byte_swap_data) |
|
|
FIELD_PREP(HAL_CE_SRC_DESC_ADDR_INFO_GATHER, 0) |
|
|
FIELD_PREP(HAL_CE_SRC_DESC_ADDR_INFO_LEN, len);
|
|
desc->meta_info = FIELD_PREP(HAL_CE_SRC_DESC_META_INFO_DATA, id);
|
|
}
|
|
|
|
void ath11k_hal_ce_dst_set_desc(void *buf, dma_addr_t paddr)
|
|
{
|
|
struct hal_ce_srng_dest_desc *desc =
|
|
(struct hal_ce_srng_dest_desc *)buf;
|
|
|
|
desc->buffer_addr_low = paddr & HAL_ADDR_LSB_REG_MASK;
|
|
desc->buffer_addr_info =
|
|
FIELD_PREP(HAL_CE_DEST_DESC_ADDR_INFO_ADDR_HI,
|
|
((u64)paddr >> HAL_ADDR_MSB_REG_SHIFT));
|
|
}
|
|
|
|
u32 ath11k_hal_ce_dst_status_get_length(void *buf)
|
|
{
|
|
struct hal_ce_srng_dst_status_desc *desc =
|
|
(struct hal_ce_srng_dst_status_desc *)buf;
|
|
u32 len;
|
|
|
|
len = FIELD_GET(HAL_CE_DST_STATUS_DESC_FLAGS_LEN, desc->flags);
|
|
desc->flags &= ~HAL_CE_DST_STATUS_DESC_FLAGS_LEN;
|
|
|
|
return len;
|
|
}
|
|
|
|
void ath11k_hal_set_link_desc_addr(struct hal_wbm_link_desc *desc, u32 cookie,
|
|
dma_addr_t paddr)
|
|
{
|
|
desc->buf_addr_info.info0 = FIELD_PREP(BUFFER_ADDR_INFO0_ADDR,
|
|
(paddr & HAL_ADDR_LSB_REG_MASK));
|
|
desc->buf_addr_info.info1 = FIELD_PREP(BUFFER_ADDR_INFO1_ADDR,
|
|
((u64)paddr >> HAL_ADDR_MSB_REG_SHIFT)) |
|
|
FIELD_PREP(BUFFER_ADDR_INFO1_RET_BUF_MGR, 1) |
|
|
FIELD_PREP(BUFFER_ADDR_INFO1_SW_COOKIE, cookie);
|
|
}
|
|
|
|
u32 *ath11k_hal_srng_dst_peek(struct ath11k_base *ab, struct hal_srng *srng)
|
|
{
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
if (srng->u.dst_ring.tp != srng->u.dst_ring.cached_hp)
|
|
return (srng->ring_base_vaddr + srng->u.dst_ring.tp);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void ath11k_hal_srng_prefetch_desc(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
u32 *desc;
|
|
|
|
/* prefetch only if desc is available */
|
|
desc = ath11k_hal_srng_dst_peek(ab, srng);
|
|
if (likely(desc)) {
|
|
dma_sync_single_for_cpu(ab->dev, virt_to_phys(desc),
|
|
(srng->entry_size * sizeof(u32)),
|
|
DMA_FROM_DEVICE);
|
|
prefetch(desc);
|
|
}
|
|
}
|
|
|
|
u32 *ath11k_hal_srng_dst_get_next_entry(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
u32 *desc;
|
|
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
if (srng->u.dst_ring.tp == srng->u.dst_ring.cached_hp)
|
|
return NULL;
|
|
|
|
desc = srng->ring_base_vaddr + srng->u.dst_ring.tp;
|
|
|
|
srng->u.dst_ring.tp += srng->entry_size;
|
|
|
|
/* wrap around to start of ring*/
|
|
if (srng->u.dst_ring.tp == srng->ring_size)
|
|
srng->u.dst_ring.tp = 0;
|
|
|
|
/* Try to prefetch the next descriptor in the ring */
|
|
if (srng->flags & HAL_SRNG_FLAGS_CACHED)
|
|
ath11k_hal_srng_prefetch_desc(ab, srng);
|
|
|
|
return desc;
|
|
}
|
|
|
|
int ath11k_hal_srng_dst_num_free(struct ath11k_base *ab, struct hal_srng *srng,
|
|
bool sync_hw_ptr)
|
|
{
|
|
u32 tp, hp;
|
|
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
tp = srng->u.dst_ring.tp;
|
|
|
|
if (sync_hw_ptr) {
|
|
hp = *srng->u.dst_ring.hp_addr;
|
|
srng->u.dst_ring.cached_hp = hp;
|
|
} else {
|
|
hp = srng->u.dst_ring.cached_hp;
|
|
}
|
|
|
|
if (hp >= tp)
|
|
return (hp - tp) / srng->entry_size;
|
|
else
|
|
return (srng->ring_size - tp + hp) / srng->entry_size;
|
|
}
|
|
|
|
/* Returns number of available entries in src ring */
|
|
int ath11k_hal_srng_src_num_free(struct ath11k_base *ab, struct hal_srng *srng,
|
|
bool sync_hw_ptr)
|
|
{
|
|
u32 tp, hp;
|
|
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
hp = srng->u.src_ring.hp;
|
|
|
|
if (sync_hw_ptr) {
|
|
tp = *srng->u.src_ring.tp_addr;
|
|
srng->u.src_ring.cached_tp = tp;
|
|
} else {
|
|
tp = srng->u.src_ring.cached_tp;
|
|
}
|
|
|
|
if (tp > hp)
|
|
return ((tp - hp) / srng->entry_size) - 1;
|
|
else
|
|
return ((srng->ring_size - hp + tp) / srng->entry_size) - 1;
|
|
}
|
|
|
|
u32 *ath11k_hal_srng_src_get_next_entry(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
u32 *desc;
|
|
u32 next_hp;
|
|
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
/* TODO: Using % is expensive, but we have to do this since size of some
|
|
* SRNG rings is not power of 2 (due to descriptor sizes). Need to see
|
|
* if separate function is defined for rings having power of 2 ring size
|
|
* (TCL2SW, REO2SW, SW2RXDMA and CE rings) so that we can avoid the
|
|
* overhead of % by using mask (with &).
|
|
*/
|
|
next_hp = (srng->u.src_ring.hp + srng->entry_size) % srng->ring_size;
|
|
|
|
if (next_hp == srng->u.src_ring.cached_tp)
|
|
return NULL;
|
|
|
|
desc = srng->ring_base_vaddr + srng->u.src_ring.hp;
|
|
srng->u.src_ring.hp = next_hp;
|
|
|
|
/* TODO: Reap functionality is not used by all rings. If particular
|
|
* ring does not use reap functionality, we need not update reap_hp
|
|
* with next_hp pointer. Need to make sure a separate function is used
|
|
* before doing any optimization by removing below code updating
|
|
* reap_hp.
|
|
*/
|
|
srng->u.src_ring.reap_hp = next_hp;
|
|
|
|
return desc;
|
|
}
|
|
|
|
u32 *ath11k_hal_srng_src_reap_next(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
u32 *desc;
|
|
u32 next_reap_hp;
|
|
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
next_reap_hp = (srng->u.src_ring.reap_hp + srng->entry_size) %
|
|
srng->ring_size;
|
|
|
|
if (next_reap_hp == srng->u.src_ring.cached_tp)
|
|
return NULL;
|
|
|
|
desc = srng->ring_base_vaddr + next_reap_hp;
|
|
srng->u.src_ring.reap_hp = next_reap_hp;
|
|
|
|
return desc;
|
|
}
|
|
|
|
u32 *ath11k_hal_srng_src_get_next_reaped(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
u32 *desc;
|
|
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
if (srng->u.src_ring.hp == srng->u.src_ring.reap_hp)
|
|
return NULL;
|
|
|
|
desc = srng->ring_base_vaddr + srng->u.src_ring.hp;
|
|
srng->u.src_ring.hp = (srng->u.src_ring.hp + srng->entry_size) %
|
|
srng->ring_size;
|
|
|
|
return desc;
|
|
}
|
|
|
|
u32 *ath11k_hal_srng_src_peek(struct ath11k_base *ab, struct hal_srng *srng)
|
|
{
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
if (((srng->u.src_ring.hp + srng->entry_size) % srng->ring_size) ==
|
|
srng->u.src_ring.cached_tp)
|
|
return NULL;
|
|
|
|
return srng->ring_base_vaddr + srng->u.src_ring.hp;
|
|
}
|
|
|
|
void ath11k_hal_srng_access_begin(struct ath11k_base *ab, struct hal_srng *srng)
|
|
{
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
if (srng->ring_dir == HAL_SRNG_DIR_SRC) {
|
|
srng->u.src_ring.cached_tp =
|
|
*(volatile u32 *)srng->u.src_ring.tp_addr;
|
|
} else {
|
|
srng->u.dst_ring.cached_hp = *srng->u.dst_ring.hp_addr;
|
|
|
|
/* Try to prefetch the next descriptor in the ring */
|
|
if (srng->flags & HAL_SRNG_FLAGS_CACHED)
|
|
ath11k_hal_srng_prefetch_desc(ab, srng);
|
|
}
|
|
}
|
|
|
|
/* Update cached ring head/tail pointers to HW. ath11k_hal_srng_access_begin()
|
|
* should have been called before this.
|
|
*/
|
|
void ath11k_hal_srng_access_end(struct ath11k_base *ab, struct hal_srng *srng)
|
|
{
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
/* TODO: See if we need a write memory barrier here */
|
|
if (srng->flags & HAL_SRNG_FLAGS_LMAC_RING) {
|
|
/* For LMAC rings, ring pointer updates are done through FW and
|
|
* hence written to a shared memory location that is read by FW
|
|
*/
|
|
if (srng->ring_dir == HAL_SRNG_DIR_SRC) {
|
|
srng->u.src_ring.last_tp =
|
|
*(volatile u32 *)srng->u.src_ring.tp_addr;
|
|
*srng->u.src_ring.hp_addr = srng->u.src_ring.hp;
|
|
} else {
|
|
srng->u.dst_ring.last_hp = *srng->u.dst_ring.hp_addr;
|
|
*srng->u.dst_ring.tp_addr = srng->u.dst_ring.tp;
|
|
}
|
|
} else {
|
|
if (srng->ring_dir == HAL_SRNG_DIR_SRC) {
|
|
srng->u.src_ring.last_tp =
|
|
*(volatile u32 *)srng->u.src_ring.tp_addr;
|
|
ath11k_hif_write32(ab,
|
|
(unsigned long)srng->u.src_ring.hp_addr -
|
|
(unsigned long)ab->mem,
|
|
srng->u.src_ring.hp);
|
|
} else {
|
|
srng->u.dst_ring.last_hp = *srng->u.dst_ring.hp_addr;
|
|
ath11k_hif_write32(ab,
|
|
(unsigned long)srng->u.dst_ring.tp_addr -
|
|
(unsigned long)ab->mem,
|
|
srng->u.dst_ring.tp);
|
|
}
|
|
}
|
|
|
|
srng->timestamp = jiffies;
|
|
}
|
|
|
|
void ath11k_hal_setup_link_idle_list(struct ath11k_base *ab,
|
|
struct hal_wbm_idle_scatter_list *sbuf,
|
|
u32 nsbufs, u32 tot_link_desc,
|
|
u32 end_offset)
|
|
{
|
|
struct ath11k_buffer_addr *link_addr;
|
|
int i;
|
|
u32 reg_scatter_buf_sz = HAL_WBM_IDLE_SCATTER_BUF_SIZE / 64;
|
|
|
|
link_addr = (void *)sbuf[0].vaddr + HAL_WBM_IDLE_SCATTER_BUF_SIZE;
|
|
|
|
for (i = 1; i < nsbufs; i++) {
|
|
link_addr->info0 = sbuf[i].paddr & HAL_ADDR_LSB_REG_MASK;
|
|
link_addr->info1 = FIELD_PREP(
|
|
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_39_32,
|
|
(u64)sbuf[i].paddr >> HAL_ADDR_MSB_REG_SHIFT) |
|
|
FIELD_PREP(
|
|
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_MATCH_TAG,
|
|
BASE_ADDR_MATCH_TAG_VAL);
|
|
|
|
link_addr = (void *)sbuf[i].vaddr +
|
|
HAL_WBM_IDLE_SCATTER_BUF_SIZE;
|
|
}
|
|
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_R0_IDLE_LIST_CONTROL_ADDR,
|
|
FIELD_PREP(HAL_WBM_SCATTER_BUFFER_SIZE, reg_scatter_buf_sz) |
|
|
FIELD_PREP(HAL_WBM_LINK_DESC_IDLE_LIST_MODE, 0x1));
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_R0_IDLE_LIST_SIZE_ADDR,
|
|
FIELD_PREP(HAL_WBM_SCATTER_RING_SIZE_OF_IDLE_LINK_DESC_LIST,
|
|
reg_scatter_buf_sz * nsbufs));
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG +
|
|
HAL_WBM_SCATTERED_RING_BASE_LSB,
|
|
FIELD_PREP(BUFFER_ADDR_INFO0_ADDR,
|
|
sbuf[0].paddr & HAL_ADDR_LSB_REG_MASK));
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG +
|
|
HAL_WBM_SCATTERED_RING_BASE_MSB,
|
|
FIELD_PREP(
|
|
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_39_32,
|
|
(u64)sbuf[0].paddr >> HAL_ADDR_MSB_REG_SHIFT) |
|
|
FIELD_PREP(
|
|
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_MATCH_TAG,
|
|
BASE_ADDR_MATCH_TAG_VAL));
|
|
|
|
/* Setup head and tail pointers for the idle list */
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG +
|
|
HAL_WBM_SCATTERED_DESC_PTR_HEAD_INFO_IX0,
|
|
FIELD_PREP(BUFFER_ADDR_INFO0_ADDR,
|
|
sbuf[nsbufs - 1].paddr));
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG +
|
|
HAL_WBM_SCATTERED_DESC_PTR_HEAD_INFO_IX1,
|
|
FIELD_PREP(
|
|
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_39_32,
|
|
((u64)sbuf[nsbufs - 1].paddr >>
|
|
HAL_ADDR_MSB_REG_SHIFT)) |
|
|
FIELD_PREP(HAL_WBM_SCATTERED_DESC_HEAD_P_OFFSET_IX1,
|
|
(end_offset >> 2)));
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG +
|
|
HAL_WBM_SCATTERED_DESC_PTR_HEAD_INFO_IX0,
|
|
FIELD_PREP(BUFFER_ADDR_INFO0_ADDR,
|
|
sbuf[0].paddr));
|
|
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG +
|
|
HAL_WBM_SCATTERED_DESC_PTR_TAIL_INFO_IX0,
|
|
FIELD_PREP(BUFFER_ADDR_INFO0_ADDR,
|
|
sbuf[0].paddr));
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG +
|
|
HAL_WBM_SCATTERED_DESC_PTR_TAIL_INFO_IX1,
|
|
FIELD_PREP(
|
|
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_39_32,
|
|
((u64)sbuf[0].paddr >> HAL_ADDR_MSB_REG_SHIFT)) |
|
|
FIELD_PREP(HAL_WBM_SCATTERED_DESC_TAIL_P_OFFSET_IX1,
|
|
0));
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG +
|
|
HAL_WBM_SCATTERED_DESC_PTR_HP_ADDR,
|
|
2 * tot_link_desc);
|
|
|
|
/* Enable the SRNG */
|
|
ath11k_hif_write32(ab,
|
|
HAL_SEQ_WCSS_UMAC_WBM_REG +
|
|
HAL_WBM_IDLE_LINK_RING_MISC_ADDR(ab), 0x40);
|
|
}
|
|
|
|
int ath11k_hal_srng_setup(struct ath11k_base *ab, enum hal_ring_type type,
|
|
int ring_num, int mac_id,
|
|
struct hal_srng_params *params)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
struct hal_srng_config *srng_config = &ab->hal.srng_config[type];
|
|
struct hal_srng *srng;
|
|
int ring_id;
|
|
u32 lmac_idx;
|
|
int i;
|
|
u32 reg_base;
|
|
|
|
ring_id = ath11k_hal_srng_get_ring_id(ab, type, ring_num, mac_id);
|
|
if (ring_id < 0)
|
|
return ring_id;
|
|
|
|
srng = &hal->srng_list[ring_id];
|
|
|
|
srng->ring_id = ring_id;
|
|
srng->ring_dir = srng_config->ring_dir;
|
|
srng->ring_base_paddr = params->ring_base_paddr;
|
|
srng->ring_base_vaddr = params->ring_base_vaddr;
|
|
srng->entry_size = srng_config->entry_size;
|
|
srng->num_entries = params->num_entries;
|
|
srng->ring_size = srng->entry_size * srng->num_entries;
|
|
srng->intr_batch_cntr_thres_entries =
|
|
params->intr_batch_cntr_thres_entries;
|
|
srng->intr_timer_thres_us = params->intr_timer_thres_us;
|
|
srng->flags = params->flags;
|
|
srng->msi_addr = params->msi_addr;
|
|
srng->msi_data = params->msi_data;
|
|
srng->initialized = 1;
|
|
spin_lock_init(&srng->lock);
|
|
lockdep_set_class(&srng->lock, hal->srng_key + ring_id);
|
|
|
|
for (i = 0; i < HAL_SRNG_NUM_REG_GRP; i++) {
|
|
srng->hwreg_base[i] = srng_config->reg_start[i] +
|
|
(ring_num * srng_config->reg_size[i]);
|
|
}
|
|
|
|
memset(srng->ring_base_vaddr, 0,
|
|
(srng->entry_size * srng->num_entries) << 2);
|
|
|
|
/* TODO: Add comments on these swap configurations */
|
|
if (IS_ENABLED(CONFIG_CPU_BIG_ENDIAN))
|
|
srng->flags |= HAL_SRNG_FLAGS_MSI_SWAP | HAL_SRNG_FLAGS_DATA_TLV_SWAP |
|
|
HAL_SRNG_FLAGS_RING_PTR_SWAP;
|
|
|
|
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R2];
|
|
|
|
if (srng->ring_dir == HAL_SRNG_DIR_SRC) {
|
|
srng->u.src_ring.hp = 0;
|
|
srng->u.src_ring.cached_tp = 0;
|
|
srng->u.src_ring.reap_hp = srng->ring_size - srng->entry_size;
|
|
srng->u.src_ring.tp_addr = (void *)(hal->rdp.vaddr + ring_id);
|
|
srng->u.src_ring.low_threshold = params->low_threshold *
|
|
srng->entry_size;
|
|
if (srng_config->lmac_ring) {
|
|
lmac_idx = ring_id - HAL_SRNG_RING_ID_LMAC1_ID_START;
|
|
srng->u.src_ring.hp_addr = (void *)(hal->wrp.vaddr +
|
|
lmac_idx);
|
|
srng->flags |= HAL_SRNG_FLAGS_LMAC_RING;
|
|
} else {
|
|
if (!ab->hw_params.supports_shadow_regs)
|
|
srng->u.src_ring.hp_addr =
|
|
(u32 *)((unsigned long)ab->mem + reg_base);
|
|
else
|
|
ath11k_dbg(ab, ATH11k_DBG_HAL,
|
|
"hal type %d ring_num %d reg_base 0x%x shadow 0x%lx\n",
|
|
type, ring_num,
|
|
reg_base,
|
|
(unsigned long)srng->u.src_ring.hp_addr -
|
|
(unsigned long)ab->mem);
|
|
}
|
|
} else {
|
|
/* During initialization loop count in all the descriptors
|
|
* will be set to zero, and HW will set it to 1 on completing
|
|
* descriptor update in first loop, and increments it by 1 on
|
|
* subsequent loops (loop count wraps around after reaching
|
|
* 0xffff). The 'loop_cnt' in SW ring state is the expected
|
|
* loop count in descriptors updated by HW (to be processed
|
|
* by SW).
|
|
*/
|
|
srng->u.dst_ring.loop_cnt = 1;
|
|
srng->u.dst_ring.tp = 0;
|
|
srng->u.dst_ring.cached_hp = 0;
|
|
srng->u.dst_ring.hp_addr = (void *)(hal->rdp.vaddr + ring_id);
|
|
if (srng_config->lmac_ring) {
|
|
/* For LMAC rings, tail pointer updates will be done
|
|
* through FW by writing to a shared memory location
|
|
*/
|
|
lmac_idx = ring_id - HAL_SRNG_RING_ID_LMAC1_ID_START;
|
|
srng->u.dst_ring.tp_addr = (void *)(hal->wrp.vaddr +
|
|
lmac_idx);
|
|
srng->flags |= HAL_SRNG_FLAGS_LMAC_RING;
|
|
} else {
|
|
if (!ab->hw_params.supports_shadow_regs)
|
|
srng->u.dst_ring.tp_addr =
|
|
(u32 *)((unsigned long)ab->mem + reg_base +
|
|
(HAL_REO1_RING_TP(ab) - HAL_REO1_RING_HP(ab)));
|
|
else
|
|
ath11k_dbg(ab, ATH11k_DBG_HAL,
|
|
"type %d ring_num %d target_reg 0x%x shadow 0x%lx\n",
|
|
type, ring_num,
|
|
reg_base + (HAL_REO1_RING_TP(ab) -
|
|
HAL_REO1_RING_HP(ab)),
|
|
(unsigned long)srng->u.dst_ring.tp_addr -
|
|
(unsigned long)ab->mem);
|
|
}
|
|
}
|
|
|
|
if (srng_config->lmac_ring)
|
|
return ring_id;
|
|
|
|
ath11k_hal_srng_hw_init(ab, srng);
|
|
|
|
if (type == HAL_CE_DST) {
|
|
srng->u.dst_ring.max_buffer_length = params->max_buffer_len;
|
|
ath11k_hal_ce_dst_setup(ab, srng, ring_num);
|
|
}
|
|
|
|
return ring_id;
|
|
}
|
|
|
|
static void ath11k_hal_srng_update_hp_tp_addr(struct ath11k_base *ab,
|
|
int shadow_cfg_idx,
|
|
enum hal_ring_type ring_type,
|
|
int ring_num)
|
|
{
|
|
struct hal_srng *srng;
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
int ring_id;
|
|
struct hal_srng_config *srng_config = &hal->srng_config[ring_type];
|
|
|
|
ring_id = ath11k_hal_srng_get_ring_id(ab, ring_type, ring_num, 0);
|
|
if (ring_id < 0)
|
|
return;
|
|
|
|
srng = &hal->srng_list[ring_id];
|
|
|
|
if (srng_config->ring_dir == HAL_SRNG_DIR_DST)
|
|
srng->u.dst_ring.tp_addr = (u32 *)(HAL_SHADOW_REG(shadow_cfg_idx) +
|
|
(unsigned long)ab->mem);
|
|
else
|
|
srng->u.src_ring.hp_addr = (u32 *)(HAL_SHADOW_REG(shadow_cfg_idx) +
|
|
(unsigned long)ab->mem);
|
|
}
|
|
|
|
int ath11k_hal_srng_update_shadow_config(struct ath11k_base *ab,
|
|
enum hal_ring_type ring_type,
|
|
int ring_num)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
struct hal_srng_config *srng_config = &hal->srng_config[ring_type];
|
|
int shadow_cfg_idx = hal->num_shadow_reg_configured;
|
|
u32 target_reg;
|
|
|
|
if (shadow_cfg_idx >= HAL_SHADOW_NUM_REGS)
|
|
return -EINVAL;
|
|
|
|
hal->num_shadow_reg_configured++;
|
|
|
|
target_reg = srng_config->reg_start[HAL_HP_OFFSET_IN_REG_START];
|
|
target_reg += srng_config->reg_size[HAL_HP_OFFSET_IN_REG_START] *
|
|
ring_num;
|
|
|
|
/* For destination ring, shadow the TP */
|
|
if (srng_config->ring_dir == HAL_SRNG_DIR_DST)
|
|
target_reg += HAL_OFFSET_FROM_HP_TO_TP;
|
|
|
|
hal->shadow_reg_addr[shadow_cfg_idx] = target_reg;
|
|
|
|
/* update hp/tp addr to hal structure*/
|
|
ath11k_hal_srng_update_hp_tp_addr(ab, shadow_cfg_idx, ring_type,
|
|
ring_num);
|
|
|
|
ath11k_dbg(ab, ATH11k_DBG_HAL,
|
|
"target_reg %x, shadow reg 0x%x shadow_idx 0x%x, ring_type %d, ring num %d",
|
|
target_reg,
|
|
HAL_SHADOW_REG(shadow_cfg_idx),
|
|
shadow_cfg_idx,
|
|
ring_type, ring_num);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ath11k_hal_srng_shadow_config(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
int ring_type, ring_num;
|
|
|
|
/* update all the non-CE srngs. */
|
|
for (ring_type = 0; ring_type < HAL_MAX_RING_TYPES; ring_type++) {
|
|
struct hal_srng_config *srng_config = &hal->srng_config[ring_type];
|
|
|
|
if (ring_type == HAL_CE_SRC ||
|
|
ring_type == HAL_CE_DST ||
|
|
ring_type == HAL_CE_DST_STATUS)
|
|
continue;
|
|
|
|
if (srng_config->lmac_ring)
|
|
continue;
|
|
|
|
for (ring_num = 0; ring_num < srng_config->max_rings; ring_num++)
|
|
ath11k_hal_srng_update_shadow_config(ab, ring_type, ring_num);
|
|
}
|
|
}
|
|
|
|
void ath11k_hal_srng_get_shadow_config(struct ath11k_base *ab,
|
|
u32 **cfg, u32 *len)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
|
|
*len = hal->num_shadow_reg_configured;
|
|
*cfg = hal->shadow_reg_addr;
|
|
}
|
|
|
|
void ath11k_hal_srng_shadow_update_hp_tp(struct ath11k_base *ab,
|
|
struct hal_srng *srng)
|
|
{
|
|
lockdep_assert_held(&srng->lock);
|
|
|
|
/* check whether the ring is emptry. Update the shadow
|
|
* HP only when then ring isn't' empty.
|
|
*/
|
|
if (srng->ring_dir == HAL_SRNG_DIR_SRC &&
|
|
*srng->u.src_ring.tp_addr != srng->u.src_ring.hp)
|
|
ath11k_hal_srng_access_end(ab, srng);
|
|
}
|
|
|
|
static int ath11k_hal_srng_create_config(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
struct hal_srng_config *s;
|
|
|
|
hal->srng_config = kmemdup(hw_srng_config_template,
|
|
sizeof(hw_srng_config_template),
|
|
GFP_KERNEL);
|
|
if (!hal->srng_config)
|
|
return -ENOMEM;
|
|
|
|
s = &hal->srng_config[HAL_REO_DST];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO1_RING_BASE_LSB(ab);
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO1_RING_HP(ab);
|
|
s->reg_size[0] = HAL_REO2_RING_BASE_LSB(ab) - HAL_REO1_RING_BASE_LSB(ab);
|
|
s->reg_size[1] = HAL_REO2_RING_HP(ab) - HAL_REO1_RING_HP(ab);
|
|
|
|
s = &hal->srng_config[HAL_REO_EXCEPTION];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_TCL_RING_BASE_LSB(ab);
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_TCL_RING_HP(ab);
|
|
|
|
s = &hal->srng_config[HAL_REO_REINJECT];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_SW2REO_RING_BASE_LSB;
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_SW2REO_RING_HP;
|
|
|
|
s = &hal->srng_config[HAL_REO_CMD];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_CMD_RING_BASE_LSB;
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_CMD_HP;
|
|
|
|
s = &hal->srng_config[HAL_REO_STATUS];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_STATUS_RING_BASE_LSB(ab);
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_STATUS_HP(ab);
|
|
|
|
s = &hal->srng_config[HAL_TCL_DATA];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL1_RING_BASE_LSB(ab);
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL1_RING_HP;
|
|
s->reg_size[0] = HAL_TCL2_RING_BASE_LSB(ab) - HAL_TCL1_RING_BASE_LSB(ab);
|
|
s->reg_size[1] = HAL_TCL2_RING_HP - HAL_TCL1_RING_HP;
|
|
|
|
s = &hal->srng_config[HAL_TCL_CMD];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_RING_BASE_LSB(ab);
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_RING_HP;
|
|
|
|
s = &hal->srng_config[HAL_TCL_STATUS];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_STATUS_RING_BASE_LSB(ab);
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_STATUS_RING_HP;
|
|
|
|
s = &hal->srng_config[HAL_CE_SRC];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_CE0_SRC_REG(ab) + HAL_CE_DST_RING_BASE_LSB;
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_CE0_SRC_REG(ab) + HAL_CE_DST_RING_HP;
|
|
s->reg_size[0] = HAL_SEQ_WCSS_UMAC_CE1_SRC_REG(ab) -
|
|
HAL_SEQ_WCSS_UMAC_CE0_SRC_REG(ab);
|
|
s->reg_size[1] = HAL_SEQ_WCSS_UMAC_CE1_SRC_REG(ab) -
|
|
HAL_SEQ_WCSS_UMAC_CE0_SRC_REG(ab);
|
|
|
|
s = &hal->srng_config[HAL_CE_DST];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG(ab) + HAL_CE_DST_RING_BASE_LSB;
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG(ab) + HAL_CE_DST_RING_HP;
|
|
s->reg_size[0] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG(ab) -
|
|
HAL_SEQ_WCSS_UMAC_CE0_DST_REG(ab);
|
|
s->reg_size[1] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG(ab) -
|
|
HAL_SEQ_WCSS_UMAC_CE0_DST_REG(ab);
|
|
|
|
s = &hal->srng_config[HAL_CE_DST_STATUS];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG(ab) +
|
|
HAL_CE_DST_STATUS_RING_BASE_LSB;
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG(ab) + HAL_CE_DST_STATUS_RING_HP;
|
|
s->reg_size[0] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG(ab) -
|
|
HAL_SEQ_WCSS_UMAC_CE0_DST_REG(ab);
|
|
s->reg_size[1] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG(ab) -
|
|
HAL_SEQ_WCSS_UMAC_CE0_DST_REG(ab);
|
|
|
|
s = &hal->srng_config[HAL_WBM_IDLE_LINK];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_IDLE_LINK_RING_BASE_LSB(ab);
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_IDLE_LINK_RING_HP;
|
|
|
|
s = &hal->srng_config[HAL_SW2WBM_RELEASE];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_RELEASE_RING_BASE_LSB(ab);
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_RELEASE_RING_HP;
|
|
|
|
s = &hal->srng_config[HAL_WBM2SW_RELEASE];
|
|
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM0_RELEASE_RING_BASE_LSB(ab);
|
|
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM0_RELEASE_RING_HP;
|
|
s->reg_size[0] = HAL_WBM1_RELEASE_RING_BASE_LSB(ab) -
|
|
HAL_WBM0_RELEASE_RING_BASE_LSB(ab);
|
|
s->reg_size[1] = HAL_WBM1_RELEASE_RING_HP - HAL_WBM0_RELEASE_RING_HP;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath11k_hal_register_srng_key(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
u32 ring_id;
|
|
|
|
for (ring_id = 0; ring_id < HAL_SRNG_RING_ID_MAX; ring_id++)
|
|
lockdep_register_key(hal->srng_key + ring_id);
|
|
}
|
|
|
|
static void ath11k_hal_unregister_srng_key(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
u32 ring_id;
|
|
|
|
for (ring_id = 0; ring_id < HAL_SRNG_RING_ID_MAX; ring_id++)
|
|
lockdep_unregister_key(hal->srng_key + ring_id);
|
|
}
|
|
|
|
int ath11k_hal_srng_init(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
int ret;
|
|
|
|
memset(hal, 0, sizeof(*hal));
|
|
|
|
ret = ath11k_hal_srng_create_config(ab);
|
|
if (ret)
|
|
goto err_hal;
|
|
|
|
ret = ath11k_hal_alloc_cont_rdp(ab);
|
|
if (ret)
|
|
goto err_hal;
|
|
|
|
ret = ath11k_hal_alloc_cont_wrp(ab);
|
|
if (ret)
|
|
goto err_free_cont_rdp;
|
|
|
|
ath11k_hal_register_srng_key(ab);
|
|
|
|
return 0;
|
|
|
|
err_free_cont_rdp:
|
|
ath11k_hal_free_cont_rdp(ab);
|
|
|
|
err_hal:
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(ath11k_hal_srng_init);
|
|
|
|
void ath11k_hal_srng_deinit(struct ath11k_base *ab)
|
|
{
|
|
struct ath11k_hal *hal = &ab->hal;
|
|
|
|
ath11k_hal_unregister_srng_key(ab);
|
|
ath11k_hal_free_cont_rdp(ab);
|
|
ath11k_hal_free_cont_wrp(ab);
|
|
kfree(hal->srng_config);
|
|
}
|
|
EXPORT_SYMBOL(ath11k_hal_srng_deinit);
|
|
|
|
void ath11k_hal_dump_srng_stats(struct ath11k_base *ab)
|
|
{
|
|
struct hal_srng *srng;
|
|
struct ath11k_ext_irq_grp *irq_grp;
|
|
struct ath11k_ce_pipe *ce_pipe;
|
|
int i;
|
|
|
|
ath11k_err(ab, "Last interrupt received for each CE:\n");
|
|
for (i = 0; i < ab->hw_params.ce_count; i++) {
|
|
ce_pipe = &ab->ce.ce_pipe[i];
|
|
|
|
if (ath11k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR)
|
|
continue;
|
|
|
|
ath11k_err(ab, "CE_id %d pipe_num %d %ums before\n",
|
|
i, ce_pipe->pipe_num,
|
|
jiffies_to_msecs(jiffies - ce_pipe->timestamp));
|
|
}
|
|
|
|
ath11k_err(ab, "\nLast interrupt received for each group:\n");
|
|
for (i = 0; i < ATH11K_EXT_IRQ_GRP_NUM_MAX; i++) {
|
|
irq_grp = &ab->ext_irq_grp[i];
|
|
ath11k_err(ab, "group_id %d %ums before\n",
|
|
irq_grp->grp_id,
|
|
jiffies_to_msecs(jiffies - irq_grp->timestamp));
|
|
}
|
|
|
|
for (i = 0; i < HAL_SRNG_RING_ID_MAX; i++) {
|
|
srng = &ab->hal.srng_list[i];
|
|
|
|
if (!srng->initialized)
|
|
continue;
|
|
|
|
if (srng->ring_dir == HAL_SRNG_DIR_SRC)
|
|
ath11k_err(ab,
|
|
"src srng id %u hp %u, reap_hp %u, cur tp %u, cached tp %u last tp %u napi processed before %ums\n",
|
|
srng->ring_id, srng->u.src_ring.hp,
|
|
srng->u.src_ring.reap_hp,
|
|
*srng->u.src_ring.tp_addr, srng->u.src_ring.cached_tp,
|
|
srng->u.src_ring.last_tp,
|
|
jiffies_to_msecs(jiffies - srng->timestamp));
|
|
else if (srng->ring_dir == HAL_SRNG_DIR_DST)
|
|
ath11k_err(ab,
|
|
"dst srng id %u tp %u, cur hp %u, cached hp %u last hp %u napi processed before %ums\n",
|
|
srng->ring_id, srng->u.dst_ring.tp,
|
|
*srng->u.dst_ring.hp_addr,
|
|
srng->u.dst_ring.cached_hp,
|
|
srng->u.dst_ring.last_hp,
|
|
jiffies_to_msecs(jiffies - srng->timestamp));
|
|
}
|
|
}
|