You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
3661 lines
124 KiB
3661 lines
124 KiB
/* Copyright (c) 2014-2020, The Linux Foundation. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 and
|
|
* only version 2 as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
*/
|
|
#include <linux/firmware.h>
|
|
#include <soc/qcom/subsystem_restart.h>
|
|
#include <soc/qcom/scm.h>
|
|
#include <linux/pm_opp.h>
|
|
#include <linux/clk/qcom.h>
|
|
|
|
#include "adreno.h"
|
|
#include "a5xx_reg.h"
|
|
#include "adreno_a5xx.h"
|
|
#include "adreno_cp_parser.h"
|
|
#include "adreno_trace.h"
|
|
#include "adreno_pm4types.h"
|
|
#include "adreno_perfcounter.h"
|
|
#include "adreno_ringbuffer.h"
|
|
#include "kgsl_sharedmem.h"
|
|
#include "kgsl_log.h"
|
|
#include "kgsl.h"
|
|
#include "kgsl_trace.h"
|
|
#include "adreno_a5xx_packets.h"
|
|
|
|
static int critical_packet_constructed;
|
|
|
|
static struct kgsl_memdesc crit_pkts;
|
|
static unsigned int crit_pkts_dwords;
|
|
static struct kgsl_memdesc crit_pkts_refbuf0;
|
|
static struct kgsl_memdesc crit_pkts_refbuf1;
|
|
static struct kgsl_memdesc crit_pkts_refbuf2;
|
|
static struct kgsl_memdesc crit_pkts_refbuf3;
|
|
|
|
static const struct adreno_vbif_data a530_vbif[] = {
|
|
{A5XX_VBIF_ROUND_ROBIN_QOS_ARB, 0x00000003},
|
|
{0, 0},
|
|
};
|
|
|
|
static const struct adreno_vbif_data a540_vbif[] = {
|
|
{A5XX_VBIF_ROUND_ROBIN_QOS_ARB, 0x00000003},
|
|
{A5XX_VBIF_GATE_OFF_WRREQ_EN, 0x00000009},
|
|
{0, 0},
|
|
};
|
|
|
|
static const struct adreno_vbif_platform a5xx_vbif_platforms[] = {
|
|
{ adreno_is_a540, a540_vbif },
|
|
{ adreno_is_a530, a530_vbif },
|
|
{ adreno_is_a512, a540_vbif },
|
|
{ adreno_is_a510, a530_vbif },
|
|
{ adreno_is_a508, a530_vbif },
|
|
{ adreno_is_a504, a530_vbif },
|
|
{ adreno_is_a505, a530_vbif },
|
|
{ adreno_is_a506, a530_vbif },
|
|
};
|
|
|
|
static void a5xx_irq_storm_worker(struct work_struct *work);
|
|
static int _read_fw2_block_header(uint32_t *header, uint32_t remain,
|
|
uint32_t id, uint32_t major, uint32_t minor);
|
|
static void a5xx_gpmu_reset(struct work_struct *work);
|
|
static int a5xx_gpmu_init(struct adreno_device *adreno_dev);
|
|
|
|
/**
|
|
* Number of times to check if the regulator enabled before
|
|
* giving up and returning failure.
|
|
*/
|
|
#define PWR_RETRY 100
|
|
|
|
/**
|
|
* Number of times to check if the GPMU firmware is initialized before
|
|
* giving up and returning failure.
|
|
*/
|
|
#define GPMU_FW_INIT_RETRY 5000
|
|
|
|
#define A530_QFPROM_RAW_PTE_ROW0_MSB 0x134
|
|
#define A530_QFPROM_RAW_PTE_ROW2_MSB 0x144
|
|
|
|
static void a530_efuse_leakage(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
unsigned int row0, row2;
|
|
unsigned int multiplier, gfx_active, leakage_pwr_on, coeff;
|
|
|
|
adreno_efuse_read_u32(adreno_dev,
|
|
A530_QFPROM_RAW_PTE_ROW0_MSB, &row0);
|
|
|
|
adreno_efuse_read_u32(adreno_dev,
|
|
A530_QFPROM_RAW_PTE_ROW2_MSB, &row2);
|
|
|
|
multiplier = (row0 >> 1) & 0x3;
|
|
gfx_active = (row2 >> 2) & 0xFF;
|
|
|
|
if (of_property_read_u32(device->pdev->dev.of_node,
|
|
"qcom,base-leakage-coefficient", &coeff))
|
|
return;
|
|
|
|
leakage_pwr_on = gfx_active * (1 << multiplier);
|
|
|
|
adreno_dev->lm_leakage = (leakage_pwr_on << 16) |
|
|
((leakage_pwr_on * coeff) / 100);
|
|
}
|
|
|
|
static void a530_efuse_speed_bin(struct adreno_device *adreno_dev)
|
|
{
|
|
unsigned int val;
|
|
unsigned int speed_bin[3];
|
|
struct kgsl_device *device = &adreno_dev->dev;
|
|
|
|
if (of_property_read_u32_array(device->pdev->dev.of_node,
|
|
"qcom,gpu-speed-bin", speed_bin, 3))
|
|
return;
|
|
|
|
adreno_efuse_read_u32(adreno_dev, speed_bin[0], &val);
|
|
|
|
adreno_dev->speed_bin = (val & speed_bin[1]) >> speed_bin[2];
|
|
}
|
|
|
|
static const struct {
|
|
int (*check)(struct adreno_device *adreno_dev);
|
|
void (*func)(struct adreno_device *adreno_dev);
|
|
} a5xx_efuse_funcs[] = {
|
|
{ adreno_is_a530, a530_efuse_leakage },
|
|
{ adreno_is_a530, a530_efuse_speed_bin },
|
|
{ adreno_is_a504, a530_efuse_speed_bin },
|
|
{ adreno_is_a505, a530_efuse_speed_bin },
|
|
{ adreno_is_a512, a530_efuse_speed_bin },
|
|
{ adreno_is_a508, a530_efuse_speed_bin },
|
|
};
|
|
|
|
static void a5xx_check_features(struct adreno_device *adreno_dev)
|
|
{
|
|
unsigned int i;
|
|
|
|
if (adreno_efuse_map(adreno_dev))
|
|
return;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(a5xx_efuse_funcs); i++) {
|
|
if (a5xx_efuse_funcs[i].check(adreno_dev))
|
|
a5xx_efuse_funcs[i].func(adreno_dev);
|
|
}
|
|
|
|
adreno_efuse_unmap(adreno_dev);
|
|
}
|
|
|
|
static void a5xx_platform_setup(struct adreno_device *adreno_dev)
|
|
{
|
|
uint64_t addr;
|
|
struct adreno_gpudev *gpudev = ADRENO_GPU_DEVICE(adreno_dev);
|
|
|
|
if (adreno_is_a504_to_a506(adreno_dev) || adreno_is_a508(adreno_dev)) {
|
|
gpudev->snapshot_data->sect_sizes->cp_meq = 32;
|
|
gpudev->snapshot_data->sect_sizes->cp_merciu = 1024;
|
|
gpudev->snapshot_data->sect_sizes->roq = 256;
|
|
|
|
/* A505 & A506 having 3 XIN ports in VBIF */
|
|
gpudev->vbif_xin_halt_ctrl0_mask =
|
|
A510_VBIF_XIN_HALT_CTRL0_MASK;
|
|
} else if (adreno_is_a510(adreno_dev)) {
|
|
gpudev->snapshot_data->sect_sizes->cp_meq = 32;
|
|
gpudev->snapshot_data->sect_sizes->cp_merciu = 32;
|
|
gpudev->snapshot_data->sect_sizes->roq = 256;
|
|
|
|
/* A510 has 3 XIN ports in VBIF */
|
|
gpudev->vbif_xin_halt_ctrl0_mask =
|
|
A510_VBIF_XIN_HALT_CTRL0_MASK;
|
|
} else if (adreno_is_a540(adreno_dev) ||
|
|
adreno_is_a512(adreno_dev)) {
|
|
gpudev->snapshot_data->sect_sizes->cp_merciu = 1024;
|
|
}
|
|
|
|
/* Calculate SP local and private mem addresses */
|
|
addr = ALIGN(ADRENO_UCHE_GMEM_BASE + adreno_dev->gmem_size, SZ_64K);
|
|
adreno_dev->sp_local_gpuaddr = addr;
|
|
adreno_dev->sp_pvt_gpuaddr = addr + SZ_64K;
|
|
|
|
/* Setup defaults that might get changed by the fuse bits */
|
|
adreno_dev->lm_leakage = A530_DEFAULT_LEAKAGE;
|
|
adreno_dev->speed_bin = 0;
|
|
|
|
if (ADRENO_FEATURE(adreno_dev, ADRENO_SPTP_PC))
|
|
set_bit(ADRENO_SPTP_PC_CTRL, &adreno_dev->pwrctrl_flag);
|
|
|
|
/* Check efuse bits for various capabilties */
|
|
a5xx_check_features(adreno_dev);
|
|
}
|
|
|
|
static void a5xx_critical_packet_destroy(struct adreno_device *adreno_dev)
|
|
{
|
|
kgsl_free_global(&adreno_dev->dev, &crit_pkts);
|
|
kgsl_free_global(&adreno_dev->dev, &crit_pkts_refbuf1);
|
|
kgsl_free_global(&adreno_dev->dev, &crit_pkts_refbuf2);
|
|
kgsl_free_global(&adreno_dev->dev, &crit_pkts_refbuf3);
|
|
|
|
kgsl_iommu_unmap_global_secure_pt_entry(KGSL_DEVICE(adreno_dev),
|
|
&crit_pkts_refbuf0);
|
|
kgsl_sharedmem_free(&crit_pkts_refbuf0);
|
|
|
|
}
|
|
|
|
static void _do_fixup(const struct adreno_critical_fixup *fixups, int count,
|
|
uint64_t *gpuaddrs, unsigned int *buffer)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < count; i++) {
|
|
buffer[fixups[i].lo_offset] =
|
|
lower_32_bits(gpuaddrs[fixups[i].buffer]) |
|
|
fixups[i].mem_offset;
|
|
|
|
buffer[fixups[i].hi_offset] =
|
|
upper_32_bits(gpuaddrs[fixups[i].buffer]);
|
|
}
|
|
}
|
|
|
|
static int a5xx_critical_packet_construct(struct adreno_device *adreno_dev)
|
|
{
|
|
|
|
unsigned int *cmds;
|
|
uint64_t gpuaddrs[CRITICAL_PACKET_MAX];
|
|
int ret;
|
|
|
|
ret = kgsl_allocate_global(&adreno_dev->dev,
|
|
&crit_pkts, PAGE_SIZE,
|
|
KGSL_MEMFLAGS_GPUREADONLY,
|
|
0, "crit_pkts");
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = kgsl_allocate_user(&adreno_dev->dev, &crit_pkts_refbuf0,
|
|
PAGE_SIZE, KGSL_MEMFLAGS_SECURE);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = kgsl_iommu_map_global_secure_pt_entry(&adreno_dev->dev,
|
|
&crit_pkts_refbuf0);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = kgsl_allocate_global(&adreno_dev->dev,
|
|
&crit_pkts_refbuf1,
|
|
PAGE_SIZE, 0, 0, "crit_pkts_refbuf1");
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = kgsl_allocate_global(&adreno_dev->dev,
|
|
&crit_pkts_refbuf2,
|
|
PAGE_SIZE, 0, 0, "crit_pkts_refbuf2");
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = kgsl_allocate_global(&adreno_dev->dev,
|
|
&crit_pkts_refbuf3,
|
|
PAGE_SIZE, 0, 0, "crit_pkts_refbuf3");
|
|
if (ret)
|
|
return ret;
|
|
|
|
cmds = crit_pkts.hostptr;
|
|
|
|
gpuaddrs[CRITICAL_PACKET0] = crit_pkts_refbuf0.gpuaddr;
|
|
gpuaddrs[CRITICAL_PACKET1] = crit_pkts_refbuf1.gpuaddr;
|
|
gpuaddrs[CRITICAL_PACKET2] = crit_pkts_refbuf2.gpuaddr;
|
|
gpuaddrs[CRITICAL_PACKET3] = crit_pkts_refbuf3.gpuaddr;
|
|
|
|
crit_pkts_dwords = ARRAY_SIZE(_a5xx_critical_pkts);
|
|
|
|
memcpy(cmds, _a5xx_critical_pkts, crit_pkts_dwords << 2);
|
|
|
|
_do_fixup(critical_pkt_fixups, ARRAY_SIZE(critical_pkt_fixups),
|
|
gpuaddrs, cmds);
|
|
|
|
cmds = crit_pkts_refbuf1.hostptr;
|
|
memcpy(cmds, _a5xx_critical_pkts_mem01,
|
|
ARRAY_SIZE(_a5xx_critical_pkts_mem01) << 2);
|
|
|
|
cmds = crit_pkts_refbuf2.hostptr;
|
|
memcpy(cmds, _a5xx_critical_pkts_mem02,
|
|
ARRAY_SIZE(_a5xx_critical_pkts_mem02) << 2);
|
|
|
|
cmds = crit_pkts_refbuf3.hostptr;
|
|
memcpy(cmds, _a5xx_critical_pkts_mem03,
|
|
ARRAY_SIZE(_a5xx_critical_pkts_mem03) << 2);
|
|
|
|
_do_fixup(critical_pkt_mem03_fixups,
|
|
ARRAY_SIZE(critical_pkt_mem03_fixups), gpuaddrs, cmds);
|
|
|
|
critical_packet_constructed = 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void a5xx_init(struct adreno_device *adreno_dev)
|
|
{
|
|
if (ADRENO_FEATURE(adreno_dev, ADRENO_GPMU))
|
|
INIT_WORK(&adreno_dev->gpmu_work, a5xx_gpmu_reset);
|
|
|
|
INIT_WORK(&adreno_dev->irq_storm_work, a5xx_irq_storm_worker);
|
|
|
|
if (ADRENO_QUIRK(adreno_dev, ADRENO_QUIRK_CRITICAL_PACKETS)) {
|
|
int ret;
|
|
|
|
ret = a5xx_critical_packet_construct(adreno_dev);
|
|
if (ret)
|
|
a5xx_critical_packet_destroy(adreno_dev);
|
|
}
|
|
|
|
a5xx_crashdump_init(adreno_dev);
|
|
}
|
|
|
|
static void a5xx_remove(struct adreno_device *adreno_dev)
|
|
{
|
|
if (ADRENO_QUIRK(adreno_dev, ADRENO_QUIRK_CRITICAL_PACKETS))
|
|
a5xx_critical_packet_destroy(adreno_dev);
|
|
}
|
|
|
|
/**
|
|
* a5xx_protect_init() - Initializes register protection on a5xx
|
|
* @device: Pointer to the device structure
|
|
* Performs register writes to enable protected access to sensitive
|
|
* registers
|
|
*/
|
|
static void a5xx_protect_init(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
int index = 0;
|
|
struct kgsl_protected_registers *iommu_regs;
|
|
|
|
/* enable access protection to privileged registers */
|
|
kgsl_regwrite(device, A5XX_CP_PROTECT_CNTL, 0x00000007);
|
|
|
|
/* RBBM registers */
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x4, 2);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x8, 3);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x10, 4);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x20, 5);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x40, 6);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x80, 6);
|
|
|
|
/* Content protection registers */
|
|
adreno_set_protected_registers(adreno_dev, &index,
|
|
A5XX_RBBM_SECVID_TSB_TRUSTED_BASE_LO, 4);
|
|
adreno_set_protected_registers(adreno_dev, &index,
|
|
A5XX_RBBM_SECVID_TRUST_CNTL, 1);
|
|
|
|
/* CP registers */
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x800, 6);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x840, 3);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x880, 5);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0x0AA0, 0);
|
|
|
|
/* RB registers */
|
|
adreno_set_protected_registers(adreno_dev, &index, 0xCC0, 0);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0xCF0, 1);
|
|
|
|
/* VPC registers */
|
|
adreno_set_protected_registers(adreno_dev, &index, 0xE68, 3);
|
|
adreno_set_protected_registers(adreno_dev, &index, 0xE70, 4);
|
|
|
|
/* UCHE registers */
|
|
adreno_set_protected_registers(adreno_dev, &index, 0xE80, ilog2(16));
|
|
|
|
/* SMMU registers */
|
|
iommu_regs = kgsl_mmu_get_prot_regs(&device->mmu);
|
|
if (iommu_regs)
|
|
adreno_set_protected_registers(adreno_dev, &index,
|
|
iommu_regs->base, ilog2(iommu_regs->range));
|
|
}
|
|
|
|
/*
|
|
* a5xx_is_sptp_idle() - A530 SP/TP/RAC should be power collapsed to be
|
|
* considered idle
|
|
* @adreno_dev: The adreno_device pointer
|
|
*/
|
|
static bool a5xx_is_sptp_idle(struct adreno_device *adreno_dev)
|
|
{
|
|
unsigned int reg;
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
/* If feature is not supported or enabled, no worry */
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_SPTP_PC) ||
|
|
!test_bit(ADRENO_SPTP_PC_CTRL, &adreno_dev->pwrctrl_flag))
|
|
return true;
|
|
kgsl_regread(device, A5XX_GPMU_SP_PWR_CLK_STATUS, ®);
|
|
if (reg & BIT(20))
|
|
return false;
|
|
kgsl_regread(device, A5XX_GPMU_RBCCU_PWR_CLK_STATUS, ®);
|
|
return !(reg & BIT(20));
|
|
}
|
|
|
|
/*
|
|
* _poll_gdsc_status() - Poll the GDSC status register
|
|
* @adreno_dev: The adreno device pointer
|
|
* @status_reg: Offset of the status register
|
|
* @status_value: The expected bit value
|
|
*
|
|
* Poll the status register till the power-on bit is equal to the
|
|
* expected value or the max retries are exceeded.
|
|
*/
|
|
static int _poll_gdsc_status(struct adreno_device *adreno_dev,
|
|
unsigned int status_reg,
|
|
unsigned int status_value)
|
|
{
|
|
unsigned int reg, retry = PWR_RETRY;
|
|
|
|
/* Bit 20 is the power on bit of SPTP and RAC GDSC status register */
|
|
do {
|
|
udelay(1);
|
|
kgsl_regread(KGSL_DEVICE(adreno_dev), status_reg, ®);
|
|
} while (((reg & BIT(20)) != (status_value << 20)) && retry--);
|
|
if ((reg & BIT(20)) != (status_value << 20))
|
|
return -ETIMEDOUT;
|
|
return 0;
|
|
}
|
|
|
|
static void a5xx_restore_isense_regs(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
unsigned int reg, i, ramp = GPMU_ISENSE_SAVE;
|
|
static unsigned int isense_regs[6] = {0xFFFF}, isense_reg_addr[] = {
|
|
A5XX_GPU_CS_DECIMAL_ALIGN,
|
|
A5XX_GPU_CS_SENSOR_PARAM_CORE_1,
|
|
A5XX_GPU_CS_SENSOR_PARAM_CORE_2,
|
|
A5XX_GPU_CS_SW_OV_FUSE_EN,
|
|
A5XX_GPU_CS_ENDPOINT_CALIBRATION_DONE,
|
|
A5XX_GPMU_TEMP_SENSOR_CONFIG};
|
|
|
|
if (!adreno_is_a540(adreno_dev))
|
|
return;
|
|
|
|
/* read signature */
|
|
kgsl_regread(device, ramp++, ®);
|
|
|
|
if (reg == 0xBABEFACE) {
|
|
/* store memory locations in buffer */
|
|
for (i = 0; i < ARRAY_SIZE(isense_regs); i++)
|
|
kgsl_regread(device, ramp + i, isense_regs + i);
|
|
|
|
/* clear signature */
|
|
kgsl_regwrite(device, GPMU_ISENSE_SAVE, 0x0);
|
|
}
|
|
|
|
/* if we never stored memory locations - do nothing */
|
|
if (isense_regs[0] == 0xFFFF)
|
|
return;
|
|
|
|
/* restore registers from memory */
|
|
for (i = 0; i < ARRAY_SIZE(isense_reg_addr); i++)
|
|
kgsl_regwrite(device, isense_reg_addr[i], isense_regs[i]);
|
|
|
|
}
|
|
|
|
/*
|
|
* a5xx_regulator_enable() - Enable any necessary HW regulators
|
|
* @adreno_dev: The adreno device pointer
|
|
*
|
|
* Some HW blocks may need their regulators explicitly enabled
|
|
* on a restart. Clocks must be on during this call.
|
|
*/
|
|
static int a5xx_regulator_enable(struct adreno_device *adreno_dev)
|
|
{
|
|
unsigned int ret;
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
if (!(adreno_is_a530(adreno_dev) || adreno_is_a540(adreno_dev))) {
|
|
/* Halt the sp_input_clk at HM level */
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL, 0x00000055);
|
|
a5xx_hwcg_set(adreno_dev, true);
|
|
/* Turn on sp_input_clk at HM level */
|
|
kgsl_regrmw(device, A5XX_RBBM_CLOCK_CNTL, 0xFF, 0);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Turn on smaller power domain first to reduce voltage droop.
|
|
* Set the default register values; set SW_COLLAPSE to 0.
|
|
*/
|
|
kgsl_regwrite(device, A5XX_GPMU_RBCCU_POWER_CNTL, 0x778000);
|
|
/* Insert a delay between RAC and SPTP GDSC to reduce voltage droop */
|
|
udelay(3);
|
|
ret = _poll_gdsc_status(adreno_dev, A5XX_GPMU_RBCCU_PWR_CLK_STATUS, 1);
|
|
if (ret) {
|
|
KGSL_PWR_ERR(device, "RBCCU GDSC enable failed\n");
|
|
return ret;
|
|
}
|
|
|
|
kgsl_regwrite(device, A5XX_GPMU_SP_POWER_CNTL, 0x778000);
|
|
ret = _poll_gdsc_status(adreno_dev, A5XX_GPMU_SP_PWR_CLK_STATUS, 1);
|
|
if (ret) {
|
|
KGSL_PWR_ERR(device, "SPTP GDSC enable failed\n");
|
|
return ret;
|
|
}
|
|
|
|
/* Disable SP clock */
|
|
kgsl_regrmw(device, A5XX_GPMU_GPMU_SP_CLOCK_CONTROL,
|
|
CNTL_IP_CLK_ENABLE, 0);
|
|
/* Enable hardware clockgating */
|
|
a5xx_hwcg_set(adreno_dev, true);
|
|
/* Enable SP clock */
|
|
kgsl_regrmw(device, A5XX_GPMU_GPMU_SP_CLOCK_CONTROL,
|
|
CNTL_IP_CLK_ENABLE, 1);
|
|
|
|
a5xx_restore_isense_regs(adreno_dev);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* a5xx_regulator_disable() - Disable any necessary HW regulators
|
|
* @adreno_dev: The adreno device pointer
|
|
*
|
|
* Some HW blocks may need their regulators explicitly disabled
|
|
* on a power down to prevent current spikes. Clocks must be on
|
|
* during this call.
|
|
*/
|
|
static void a5xx_regulator_disable(struct adreno_device *adreno_dev)
|
|
{
|
|
unsigned int reg;
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
if (adreno_is_a512(adreno_dev) || adreno_is_a508(adreno_dev))
|
|
return;
|
|
|
|
/* If feature is not supported or not enabled */
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_SPTP_PC) ||
|
|
!test_bit(ADRENO_SPTP_PC_CTRL, &adreno_dev->pwrctrl_flag)) {
|
|
/* Set the default register values; set SW_COLLAPSE to 1 */
|
|
kgsl_regwrite(device, A5XX_GPMU_SP_POWER_CNTL, 0x778001);
|
|
/*
|
|
* Insert a delay between SPTP and RAC GDSC to reduce voltage
|
|
* droop.
|
|
*/
|
|
udelay(3);
|
|
if (_poll_gdsc_status(adreno_dev,
|
|
A5XX_GPMU_SP_PWR_CLK_STATUS, 0))
|
|
KGSL_PWR_WARN(device, "SPTP GDSC disable failed\n");
|
|
|
|
kgsl_regwrite(device, A5XX_GPMU_RBCCU_POWER_CNTL, 0x778001);
|
|
if (_poll_gdsc_status(adreno_dev,
|
|
A5XX_GPMU_RBCCU_PWR_CLK_STATUS, 0))
|
|
KGSL_PWR_WARN(device, "RBCCU GDSC disable failed\n");
|
|
} else if (test_bit(ADRENO_DEVICE_GPMU_INITIALIZED,
|
|
&adreno_dev->priv)) {
|
|
/* GPMU firmware is supposed to turn off SPTP & RAC GDSCs. */
|
|
kgsl_regread(device, A5XX_GPMU_SP_PWR_CLK_STATUS, ®);
|
|
if (reg & BIT(20))
|
|
KGSL_PWR_WARN(device, "SPTP GDSC is not disabled\n");
|
|
kgsl_regread(device, A5XX_GPMU_RBCCU_PWR_CLK_STATUS, ®);
|
|
if (reg & BIT(20))
|
|
KGSL_PWR_WARN(device, "RBCCU GDSC is not disabled\n");
|
|
/*
|
|
* GPMU firmware is supposed to set GMEM to non-retention.
|
|
* Bit 14 is the memory core force on bit.
|
|
*/
|
|
kgsl_regread(device, A5XX_GPMU_RBCCU_CLOCK_CNTL, ®);
|
|
if (reg & BIT(14))
|
|
KGSL_PWR_WARN(device, "GMEM is forced on\n");
|
|
}
|
|
|
|
if (adreno_is_a530(adreno_dev)) {
|
|
/* Reset VBIF before PC to avoid popping bogus FIFO entries */
|
|
kgsl_regwrite(device, A5XX_RBBM_BLOCK_SW_RESET_CMD,
|
|
0x003C0000);
|
|
kgsl_regwrite(device, A5XX_RBBM_BLOCK_SW_RESET_CMD, 0);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* a5xx_enable_pc() - Enable the GPMU based power collapse of the SPTP and RAC
|
|
* blocks
|
|
* @adreno_dev: The adreno device pointer
|
|
*/
|
|
static void a5xx_enable_pc(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_SPTP_PC) ||
|
|
!test_bit(ADRENO_SPTP_PC_CTRL, &adreno_dev->pwrctrl_flag))
|
|
return;
|
|
|
|
kgsl_regwrite(device, A5XX_GPMU_PWR_COL_INTER_FRAME_CTRL, 0x0000007F);
|
|
kgsl_regwrite(device, A5XX_GPMU_PWR_COL_BINNING_CTRL, 0);
|
|
kgsl_regwrite(device, A5XX_GPMU_PWR_COL_INTER_FRAME_HYST, 0x000A0080);
|
|
kgsl_regwrite(device, A5XX_GPMU_PWR_COL_STAGGER_DELAY, 0x00600040);
|
|
|
|
trace_adreno_sp_tp((unsigned long) __builtin_return_address(0));
|
|
};
|
|
|
|
/*
|
|
* The maximum payload of a type4 packet is the max size minus one for the
|
|
* opcode
|
|
*/
|
|
#define TYPE4_MAX_PAYLOAD (PM4_TYPE4_PKT_SIZE_MAX - 1)
|
|
|
|
static int _gpmu_create_load_cmds(struct adreno_device *adreno_dev,
|
|
uint32_t *ucode, uint32_t size)
|
|
{
|
|
uint32_t *start, *cmds;
|
|
uint32_t offset = 0;
|
|
uint32_t cmds_size = size;
|
|
|
|
/* Add a dword for each PM4 packet */
|
|
cmds_size += (size / TYPE4_MAX_PAYLOAD) + 1;
|
|
|
|
/* Add 4 dwords for the protected mode */
|
|
cmds_size += 4;
|
|
|
|
if (adreno_dev->gpmu_cmds != NULL)
|
|
return 0;
|
|
|
|
adreno_dev->gpmu_cmds = kmalloc(cmds_size << 2, GFP_KERNEL);
|
|
if (adreno_dev->gpmu_cmds == NULL)
|
|
return -ENOMEM;
|
|
|
|
cmds = adreno_dev->gpmu_cmds;
|
|
start = cmds;
|
|
|
|
/* Turn CP protection OFF */
|
|
*cmds++ = cp_type7_packet(CP_SET_PROTECTED_MODE, 1);
|
|
*cmds++ = 0;
|
|
|
|
/*
|
|
* Prebuild the cmd stream to send to the GPU to load
|
|
* the GPMU firmware
|
|
*/
|
|
while (size > 0) {
|
|
int tmp_size = size;
|
|
|
|
if (size >= TYPE4_MAX_PAYLOAD)
|
|
tmp_size = TYPE4_MAX_PAYLOAD;
|
|
|
|
*cmds++ = cp_type4_packet(
|
|
A5XX_GPMU_INST_RAM_BASE + offset,
|
|
tmp_size);
|
|
|
|
memcpy(cmds, &ucode[offset], tmp_size << 2);
|
|
|
|
cmds += tmp_size;
|
|
offset += tmp_size;
|
|
size -= tmp_size;
|
|
}
|
|
|
|
/* Turn CP protection ON */
|
|
*cmds++ = cp_type7_packet(CP_SET_PROTECTED_MODE, 1);
|
|
*cmds++ = 1;
|
|
|
|
adreno_dev->gpmu_cmds_size = (size_t) (cmds - start);
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/*
|
|
* _load_gpmu_firmware() - Load the ucode into the GPMU RAM
|
|
* @adreno_dev: Pointer to adreno device
|
|
*/
|
|
static int _load_gpmu_firmware(struct adreno_device *adreno_dev)
|
|
{
|
|
uint32_t *data;
|
|
const struct firmware *fw = NULL;
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
const struct adreno_gpu_core *gpucore = adreno_dev->gpucore;
|
|
uint32_t *cmds, cmd_size;
|
|
int ret = -EINVAL;
|
|
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_GPMU))
|
|
return 0;
|
|
|
|
/* gpmu fw already saved and verified so do nothing new */
|
|
if (adreno_dev->gpmu_cmds_size != 0)
|
|
return 0;
|
|
|
|
if (gpucore->gpmufw_name == NULL)
|
|
return 0;
|
|
|
|
ret = request_firmware(&fw, gpucore->gpmufw_name, device->dev);
|
|
if (ret || fw == NULL) {
|
|
KGSL_CORE_ERR("request_firmware (%s) failed: %d\n",
|
|
gpucore->gpmufw_name, ret);
|
|
return ret;
|
|
}
|
|
|
|
data = (uint32_t *)fw->data;
|
|
|
|
if (data[0] >= (fw->size / sizeof(uint32_t)) || data[0] < 2)
|
|
goto err;
|
|
|
|
if (data[1] != GPMU_FIRMWARE_ID)
|
|
goto err;
|
|
ret = _read_fw2_block_header(&data[2],
|
|
data[0] - 2,
|
|
GPMU_FIRMWARE_ID,
|
|
adreno_dev->gpucore->gpmu_major,
|
|
adreno_dev->gpucore->gpmu_minor);
|
|
if (ret)
|
|
goto err;
|
|
|
|
/* Integer overflow check for cmd_size */
|
|
if (data[2] > (data[0] - 2))
|
|
goto err;
|
|
|
|
cmds = data + data[2] + 3;
|
|
cmd_size = data[0] - data[2] - 2;
|
|
|
|
if (cmd_size > GPMU_INST_RAM_SIZE) {
|
|
KGSL_CORE_ERR(
|
|
"GPMU firmware block size is larger than RAM size\n");
|
|
goto err;
|
|
}
|
|
|
|
/* Everything is cool, so create some commands */
|
|
ret = _gpmu_create_load_cmds(adreno_dev, cmds, cmd_size);
|
|
err:
|
|
if (fw)
|
|
release_firmware(fw);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int _gpmu_send_init_cmds(struct adreno_device *adreno_dev)
|
|
{
|
|
struct adreno_ringbuffer *rb = adreno_dev->cur_rb;
|
|
uint32_t *cmds;
|
|
uint32_t size = adreno_dev->gpmu_cmds_size;
|
|
int ret;
|
|
|
|
if (size == 0 || adreno_dev->gpmu_cmds == NULL)
|
|
return -EINVAL;
|
|
|
|
cmds = adreno_ringbuffer_allocspace(rb, size);
|
|
if (IS_ERR(cmds))
|
|
return PTR_ERR(cmds);
|
|
if (cmds == NULL)
|
|
return -ENOSPC;
|
|
|
|
/* Copy to the RB the predefined fw sequence cmds */
|
|
memcpy(cmds, adreno_dev->gpmu_cmds, size << 2);
|
|
|
|
ret = adreno_ringbuffer_submit_spin(rb, NULL, 2000);
|
|
if (ret != 0)
|
|
adreno_spin_idle_debug(adreno_dev,
|
|
"gpmu initialization failed to idle\n");
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* a5xx_gpmu_start() - Initialize and start the GPMU
|
|
* @adreno_dev: Pointer to adreno device
|
|
*
|
|
* Load the GPMU microcode, set up any features such as hardware clock gating
|
|
* or IFPC, and take the GPMU out of reset.
|
|
*/
|
|
static int a5xx_gpmu_start(struct adreno_device *adreno_dev)
|
|
{
|
|
int ret;
|
|
unsigned int reg, retry = GPMU_FW_INIT_RETRY;
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_GPMU))
|
|
return 0;
|
|
|
|
ret = _gpmu_send_init_cmds(adreno_dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (adreno_is_a530(adreno_dev)) {
|
|
/* GPMU clock gating setup */
|
|
kgsl_regwrite(device, A5XX_GPMU_WFI_CONFIG, 0x00004014);
|
|
}
|
|
/* Kick off GPMU firmware */
|
|
kgsl_regwrite(device, A5XX_GPMU_CM3_SYSRESET, 0);
|
|
/*
|
|
* The hardware team's estimation of GPMU firmware initialization
|
|
* latency is about 3000 cycles, that's about 5 to 24 usec.
|
|
*/
|
|
do {
|
|
udelay(1);
|
|
kgsl_regread(device, A5XX_GPMU_GENERAL_0, ®);
|
|
} while ((reg != 0xBABEFACE) && retry--);
|
|
|
|
if (reg != 0xBABEFACE) {
|
|
KGSL_CORE_ERR("GPMU firmware initialization timed out\n");
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
if (!adreno_is_a530(adreno_dev)) {
|
|
kgsl_regread(device, A5XX_GPMU_GENERAL_1, ®);
|
|
|
|
if (reg) {
|
|
KGSL_CORE_ERR(
|
|
"GPMU firmware initialization failed: %d\n",
|
|
reg);
|
|
return -EIO;
|
|
}
|
|
}
|
|
set_bit(ADRENO_DEVICE_GPMU_INITIALIZED, &adreno_dev->priv);
|
|
/*
|
|
* We are in AWARE state and IRQ line from GPU to host is
|
|
* disabled.
|
|
* Read pending GPMU interrupts and clear GPMU_RBBM_INTR_INFO.
|
|
*/
|
|
kgsl_regread(device, A5XX_GPMU_RBBM_INTR_INFO, ®);
|
|
/*
|
|
* Clear RBBM interrupt mask if any of GPMU interrupts
|
|
* are pending.
|
|
*/
|
|
if (reg)
|
|
kgsl_regwrite(device,
|
|
A5XX_RBBM_INT_CLEAR_CMD,
|
|
1 << A5XX_INT_GPMU_FIRMWARE);
|
|
return ret;
|
|
}
|
|
|
|
struct kgsl_hwcg_reg {
|
|
unsigned int off;
|
|
unsigned int val;
|
|
};
|
|
|
|
static const struct kgsl_hwcg_reg a50x_hwcg_regs[] = {
|
|
{A5XX_RBBM_CLOCK_CNTL_SP0, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP0, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_HYST_SP0, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP0, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP0, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP0, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP0, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_CNTL2_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL4_UCHE, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_HYST_UCHE, 0x00FFFFF4},
|
|
{A5XX_RBBM_CLOCK_DELAY_UCHE, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB0, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU0, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_RAC, 0x05522222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RAC, 0x00505555},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU0, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RAC, 0x07444044},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_0, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RAC, 0x00010011},
|
|
{A5XX_RBBM_CLOCK_CNTL_TSE_RAS_RBBM, 0x04222222},
|
|
{A5XX_RBBM_CLOCK_MODE_GPC, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_MODE_VFD, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_HYST_GPC, 0x04104004},
|
|
{A5XX_RBBM_CLOCK_HYST_VFD, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_HLSQ, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM, 0x00004000},
|
|
{A5XX_RBBM_CLOCK_DELAY_GPC, 0x00000200},
|
|
{A5XX_RBBM_CLOCK_DELAY_VFD, 0x00002222}
|
|
};
|
|
|
|
static const struct kgsl_hwcg_reg a510_hwcg_regs[] = {
|
|
{A5XX_RBBM_CLOCK_CNTL_SP0, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_SP1, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP0, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP1, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_HYST_SP0, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_HYST_SP1, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP0, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP1, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP0, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP1, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST_TP1, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP1, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP0, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP1, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP1, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP1, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP0, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP1, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_CNTL_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL4_UCHE, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_HYST_UCHE, 0x00444444},
|
|
{A5XX_RBBM_CLOCK_DELAY_UCHE, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB0, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB1, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU0, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU1, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_RAC, 0x05522222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RAC, 0x00505555},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU0, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU1, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RAC, 0x07444044},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_0, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_1, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RAC, 0x00010011},
|
|
{A5XX_RBBM_CLOCK_CNTL_TSE_RAS_RBBM, 0x04222222},
|
|
{A5XX_RBBM_CLOCK_MODE_GPC, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_MODE_VFD, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_HYST_GPC, 0x04104004},
|
|
{A5XX_RBBM_CLOCK_HYST_VFD, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_HLSQ, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM, 0x00004000},
|
|
{A5XX_RBBM_CLOCK_DELAY_GPC, 0x00000200},
|
|
{A5XX_RBBM_CLOCK_DELAY_VFD, 0x00002222}
|
|
};
|
|
|
|
static const struct kgsl_hwcg_reg a530_hwcg_regs[] = {
|
|
{A5XX_RBBM_CLOCK_CNTL_SP0, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_SP1, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_SP2, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_SP3, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP0, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP1, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP2, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP3, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_HYST_SP0, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_HYST_SP1, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_HYST_SP2, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_HYST_SP3, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP0, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP1, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP2, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP3, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP2, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP3, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP2, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP3, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP0, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP1, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP2, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP3, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST_TP1, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST_TP2, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST_TP3, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP1, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP2, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP3, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP0, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP1, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP2, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP3, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP1, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP2, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP3, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP1, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP2, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP3, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP0, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP1, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP2, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP3, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_CNTL_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL4_UCHE, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_HYST_UCHE, 0x00444444},
|
|
{A5XX_RBBM_CLOCK_DELAY_UCHE, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB2, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB3, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB0, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB1, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB2, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB3, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU0, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU1, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU2, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU3, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_RAC, 0x05522222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RAC, 0x00505555},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU0, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU1, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU2, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU3, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RAC, 0x07444044},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_0, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_1, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_2, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_3, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RAC, 0x00010011},
|
|
{A5XX_RBBM_CLOCK_CNTL_TSE_RAS_RBBM, 0x04222222},
|
|
{A5XX_RBBM_CLOCK_MODE_GPC, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_MODE_VFD, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_HYST_GPC, 0x04104004},
|
|
{A5XX_RBBM_CLOCK_HYST_VFD, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_HLSQ, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM, 0x00004000},
|
|
{A5XX_RBBM_CLOCK_DELAY_GPC, 0x00000200},
|
|
{A5XX_RBBM_CLOCK_DELAY_VFD, 0x00002222}
|
|
};
|
|
|
|
|
|
static const struct kgsl_hwcg_reg a540_hwcg_regs[] = {
|
|
{A5XX_RBBM_CLOCK_CNTL_SP0, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_SP1, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_SP2, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_SP3, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP0, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP1, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP2, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP3, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_HYST_SP0, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_HYST_SP1, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_HYST_SP2, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_HYST_SP3, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP0, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP1, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP2, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP3, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP2, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP3, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP2, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP3, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP0, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP1, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP2, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP3, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST_TP1, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST_TP2, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST_TP3, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP1, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP2, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP3, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP0, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP1, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP2, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP3, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP1, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP2, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP3, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP1, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP2, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP3, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP0, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP1, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP2, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP3, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_CNTL_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL4_UCHE, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_HYST_UCHE, 0x00444444},
|
|
{A5XX_RBBM_CLOCK_DELAY_UCHE, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB2, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB3, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB0, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB1, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB2, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB3, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU0, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU1, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU2, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU3, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_RAC, 0x05522222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RAC, 0x00505555},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU0, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU1, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU2, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU3, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RAC, 0x07444044},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_0, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_1, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_2, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_3, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RAC, 0x00010011},
|
|
{A5XX_RBBM_CLOCK_CNTL_TSE_RAS_RBBM, 0x04222222},
|
|
{A5XX_RBBM_CLOCK_MODE_GPC, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_MODE_VFD, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_HYST_GPC, 0x04104004},
|
|
{A5XX_RBBM_CLOCK_HYST_VFD, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_HLSQ, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM, 0x00004000},
|
|
{A5XX_RBBM_CLOCK_DELAY_GPC, 0x00000200},
|
|
{A5XX_RBBM_CLOCK_DELAY_VFD, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_GPMU, 0x00000222},
|
|
{A5XX_RBBM_CLOCK_DELAY_GPMU, 0x00000770},
|
|
{A5XX_RBBM_CLOCK_HYST_GPMU, 0x00000004}
|
|
};
|
|
|
|
static const struct kgsl_hwcg_reg a512_hwcg_regs[] = {
|
|
{A5XX_RBBM_CLOCK_CNTL_SP0, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_SP1, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP0, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_CNTL2_SP1, 0x02222220},
|
|
{A5XX_RBBM_CLOCK_HYST_SP0, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_HYST_SP1, 0x0000F3CF},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP0, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_DELAY_SP1, 0x00000080},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_TP1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_TP1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP0, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_TP1, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST_TP1, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP0, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST2_TP1, 0x77777777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP0, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_HYST3_TP1, 0x00007777},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY_TP1, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP0, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY2_TP1, 0x11111111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP0, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_DELAY3_TP1, 0x00001111},
|
|
{A5XX_RBBM_CLOCK_CNTL_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL3_UCHE, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL4_UCHE, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_HYST_UCHE, 0x00444444},
|
|
{A5XX_RBBM_CLOCK_DELAY_UCHE, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB0, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_RB1, 0x22222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB0, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RB1, 0x00222222},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU0, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_CCU1, 0x00022220},
|
|
{A5XX_RBBM_CLOCK_CNTL_RAC, 0x05522222},
|
|
{A5XX_RBBM_CLOCK_CNTL2_RAC, 0x00505555},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU0, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RB_CCU1, 0x04040404},
|
|
{A5XX_RBBM_CLOCK_HYST_RAC, 0x07444044},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_0, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_1, 0x00000002},
|
|
{A5XX_RBBM_CLOCK_DELAY_RAC, 0x00010011},
|
|
{A5XX_RBBM_CLOCK_CNTL_TSE_RAS_RBBM, 0x04222222},
|
|
{A5XX_RBBM_CLOCK_MODE_GPC, 0x02222222},
|
|
{A5XX_RBBM_CLOCK_MODE_VFD, 0x00002222},
|
|
{A5XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_HYST_GPC, 0x04104004},
|
|
{A5XX_RBBM_CLOCK_HYST_VFD, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_HLSQ, 0x00000000},
|
|
{A5XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM, 0x00004000},
|
|
{A5XX_RBBM_CLOCK_DELAY_GPC, 0x00000200},
|
|
{A5XX_RBBM_CLOCK_DELAY_VFD, 0x00002222},
|
|
};
|
|
|
|
static const struct {
|
|
int (*devfunc)(struct adreno_device *adreno_dev);
|
|
const struct kgsl_hwcg_reg *regs;
|
|
unsigned int count;
|
|
} a5xx_hwcg_registers[] = {
|
|
{ adreno_is_a540, a540_hwcg_regs, ARRAY_SIZE(a540_hwcg_regs) },
|
|
{ adreno_is_a530, a530_hwcg_regs, ARRAY_SIZE(a530_hwcg_regs) },
|
|
{ adreno_is_a512, a512_hwcg_regs, ARRAY_SIZE(a512_hwcg_regs) },
|
|
{ adreno_is_a510, a510_hwcg_regs, ARRAY_SIZE(a510_hwcg_regs) },
|
|
{ adreno_is_a504, a50x_hwcg_regs, ARRAY_SIZE(a50x_hwcg_regs) },
|
|
{ adreno_is_a505, a50x_hwcg_regs, ARRAY_SIZE(a50x_hwcg_regs) },
|
|
{ adreno_is_a506, a50x_hwcg_regs, ARRAY_SIZE(a50x_hwcg_regs) },
|
|
{ adreno_is_a508, a50x_hwcg_regs, ARRAY_SIZE(a50x_hwcg_regs) },
|
|
};
|
|
|
|
void a5xx_hwcg_set(struct adreno_device *adreno_dev, bool on)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
const struct kgsl_hwcg_reg *regs;
|
|
int i, j;
|
|
|
|
if (!test_bit(ADRENO_HWCG_CTRL, &adreno_dev->pwrctrl_flag))
|
|
return;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(a5xx_hwcg_registers); i++) {
|
|
if (a5xx_hwcg_registers[i].devfunc(adreno_dev))
|
|
break;
|
|
}
|
|
|
|
if (i == ARRAY_SIZE(a5xx_hwcg_registers))
|
|
return;
|
|
|
|
regs = a5xx_hwcg_registers[i].regs;
|
|
|
|
for (j = 0; j < a5xx_hwcg_registers[i].count; j++)
|
|
kgsl_regwrite(device, regs[j].off, on ? regs[j].val : 0);
|
|
|
|
/* enable top level HWCG */
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL, on ? 0xAAA8AA00 : 0);
|
|
kgsl_regwrite(device, A5XX_RBBM_ISDB_CNT, on ? 0x00000182 : 0x00000180);
|
|
}
|
|
|
|
static int _read_fw2_block_header(uint32_t *header, uint32_t remain,
|
|
uint32_t id, uint32_t major, uint32_t minor)
|
|
{
|
|
uint32_t header_size;
|
|
int i = 1;
|
|
|
|
if (header == NULL)
|
|
return -ENOMEM;
|
|
|
|
header_size = header[0];
|
|
/* Headers have limited size and always occur as pairs of words */
|
|
if (header_size > MAX_HEADER_SIZE || header_size >= remain ||
|
|
header_size % 2 || header_size == 0)
|
|
return -EINVAL;
|
|
/* Sequences must have an identifying id first thing in their header */
|
|
if (id == GPMU_SEQUENCE_ID) {
|
|
if (header[i] != HEADER_SEQUENCE ||
|
|
(header[i + 1] >= MAX_SEQUENCE_ID))
|
|
return -EINVAL;
|
|
i += 2;
|
|
}
|
|
for (; i < header_size; i += 2) {
|
|
switch (header[i]) {
|
|
/* Major Version */
|
|
case HEADER_MAJOR:
|
|
if ((major > header[i + 1]) &&
|
|
header[i + 1]) {
|
|
KGSL_CORE_ERR(
|
|
"GPMU major version mis-match %d, %d\n",
|
|
major, header[i + 1]);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
case HEADER_MINOR:
|
|
if (minor > header[i + 1])
|
|
KGSL_CORE_ERR(
|
|
"GPMU minor version mis-match %d %d\n",
|
|
minor, header[i + 1]);
|
|
break;
|
|
case HEADER_DATE:
|
|
case HEADER_TIME:
|
|
break;
|
|
default:
|
|
KGSL_CORE_ERR("GPMU unknown header ID %d\n",
|
|
header[i]);
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Read in the register sequence file and save pointers to the
|
|
* necessary sequences.
|
|
*
|
|
* GPU sequence file format (one dword per field unless noted):
|
|
* Block 1 length (length dword field not inclusive)
|
|
* Block 1 type = Sequence = 3
|
|
* Block Header length (length dword field not inclusive)
|
|
* BH field ID = Sequence field ID
|
|
* BH field data = Sequence ID
|
|
* BH field ID
|
|
* BH field data
|
|
* ...
|
|
* Opcode 0 ID
|
|
* Opcode 0 data M words
|
|
* Opcode 1 ID
|
|
* Opcode 1 data N words
|
|
* ...
|
|
* Opcode X ID
|
|
* Opcode X data O words
|
|
* Block 2 length...
|
|
*/
|
|
static void _load_regfile(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
const struct firmware *fw;
|
|
uint64_t block_size = 0, block_total = 0;
|
|
uint32_t fw_size, *block;
|
|
int ret = -EINVAL;
|
|
|
|
if (!adreno_dev->gpucore->regfw_name)
|
|
return;
|
|
|
|
ret = request_firmware(&fw, adreno_dev->gpucore->regfw_name,
|
|
device->dev);
|
|
if (ret) {
|
|
KGSL_PWR_ERR(device, "request firmware failed %d, %s\n",
|
|
ret, adreno_dev->gpucore->regfw_name);
|
|
return;
|
|
}
|
|
|
|
fw_size = fw->size / sizeof(uint32_t);
|
|
/* Min valid file of size 6, see file description */
|
|
if (fw_size < 6)
|
|
goto err;
|
|
block = (uint32_t *)fw->data;
|
|
/* All offset numbers calculated from file description */
|
|
while (block_total < fw_size) {
|
|
block_size = block[0];
|
|
if (((block_total + block_size) >= fw_size)
|
|
|| block_size < 5)
|
|
goto err;
|
|
if (block[1] != GPMU_SEQUENCE_ID)
|
|
goto err;
|
|
|
|
/* For now ignore blocks other than the LM sequence */
|
|
if (block[4] == LM_SEQUENCE_ID) {
|
|
ret = _read_fw2_block_header(&block[2],
|
|
block_size - 2,
|
|
GPMU_SEQUENCE_ID,
|
|
adreno_dev->gpucore->lm_major,
|
|
adreno_dev->gpucore->lm_minor);
|
|
if (ret)
|
|
goto err;
|
|
|
|
adreno_dev->lm_fw = fw;
|
|
|
|
if (block[2] > (block_size - 2))
|
|
goto err;
|
|
adreno_dev->lm_sequence = block + block[2] + 3;
|
|
adreno_dev->lm_size = block_size - block[2] - 2;
|
|
}
|
|
block_total += (block_size + 1);
|
|
block += (block_size + 1);
|
|
}
|
|
if (adreno_dev->lm_sequence)
|
|
return;
|
|
|
|
err:
|
|
release_firmware(fw);
|
|
KGSL_PWR_ERR(device,
|
|
"Register file failed to load sz=%d bsz=%llu header=%d\n",
|
|
fw_size, block_size, ret);
|
|
}
|
|
|
|
static int _execute_reg_sequence(struct adreno_device *adreno_dev,
|
|
uint32_t *opcode, uint32_t length)
|
|
{
|
|
uint32_t *cur = opcode;
|
|
uint64_t reg, val;
|
|
|
|
/* todo double check the reg writes */
|
|
while ((cur - opcode) < length) {
|
|
if (cur[0] == 1 && (length - (cur - opcode) >= 4)) {
|
|
/* Write a 32 bit value to a 64 bit reg */
|
|
reg = cur[2];
|
|
reg = (reg << 32) | cur[1];
|
|
kgsl_regwrite(KGSL_DEVICE(adreno_dev), reg, cur[3]);
|
|
cur += 4;
|
|
} else if (cur[0] == 2 && (length - (cur - opcode) >= 5)) {
|
|
/* Write a 64 bit value to a 64 bit reg */
|
|
reg = cur[2];
|
|
reg = (reg << 32) | cur[1];
|
|
val = cur[4];
|
|
val = (val << 32) | cur[3];
|
|
kgsl_regwrite(KGSL_DEVICE(adreno_dev), reg, val);
|
|
cur += 5;
|
|
} else if (cur[0] == 3 && (length - (cur - opcode) >= 2)) {
|
|
/* Delay for X usec */
|
|
udelay(cur[1]);
|
|
cur += 2;
|
|
} else
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static uint32_t _write_voltage_table(struct adreno_device *adreno_dev,
|
|
unsigned int addr)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
struct kgsl_pwrctrl *pwr = &device->pwrctrl;
|
|
int i;
|
|
struct dev_pm_opp *opp;
|
|
int levels = pwr->num_pwrlevels - 1;
|
|
unsigned int mvolt = 0;
|
|
|
|
kgsl_regwrite(device, addr++, adreno_dev->gpucore->max_power);
|
|
kgsl_regwrite(device, addr++, levels);
|
|
|
|
/* Write voltage in mV and frequency in MHz */
|
|
for (i = 0; i < levels; i++) {
|
|
opp = dev_pm_opp_find_freq_exact(&device->pdev->dev,
|
|
pwr->pwrlevels[i].gpu_freq, true);
|
|
/* _opp_get returns uV, convert to mV */
|
|
if (!IS_ERR(opp)) {
|
|
mvolt = dev_pm_opp_get_voltage(opp) / 1000;
|
|
dev_pm_opp_put(opp);
|
|
}
|
|
kgsl_regwrite(device, addr++, mvolt);
|
|
kgsl_regwrite(device, addr++,
|
|
pwr->pwrlevels[i].gpu_freq / 1000000);
|
|
}
|
|
return (levels * 2 + 2);
|
|
}
|
|
|
|
static uint32_t lm_limit(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
if (adreno_dev->lm_limit)
|
|
return adreno_dev->lm_limit;
|
|
|
|
if (of_property_read_u32(device->pdev->dev.of_node, "qcom,lm-limit",
|
|
&adreno_dev->lm_limit))
|
|
adreno_dev->lm_limit = LM_DEFAULT_LIMIT;
|
|
|
|
return adreno_dev->lm_limit;
|
|
}
|
|
/*
|
|
* a5xx_lm_init() - Initialize LM/DPM on the GPMU
|
|
* @adreno_dev: The adreno device pointer
|
|
*/
|
|
static void a530_lm_init(struct adreno_device *adreno_dev)
|
|
{
|
|
uint32_t length;
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_LM) ||
|
|
!test_bit(ADRENO_LM_CTRL, &adreno_dev->pwrctrl_flag))
|
|
return;
|
|
|
|
/* If something was wrong with the sequence file, return */
|
|
if (adreno_dev->lm_sequence == NULL)
|
|
return;
|
|
|
|
/* Write LM registers including DPM ucode, coefficients, and config */
|
|
if (_execute_reg_sequence(adreno_dev, adreno_dev->lm_sequence,
|
|
adreno_dev->lm_size)) {
|
|
/* If the sequence is invalid, it's not getting better */
|
|
adreno_dev->lm_sequence = NULL;
|
|
KGSL_PWR_WARN(device,
|
|
"Invalid LM sequence\n");
|
|
return;
|
|
}
|
|
|
|
kgsl_regwrite(device, A5XX_GPMU_TEMP_SENSOR_ID,
|
|
adreno_dev->gpucore->gpmu_tsens);
|
|
kgsl_regwrite(device, A5XX_GPMU_DELTA_TEMP_THRESHOLD, 0x1);
|
|
kgsl_regwrite(device, A5XX_GPMU_TEMP_SENSOR_CONFIG, 0x1);
|
|
|
|
kgsl_regwrite(device, A5XX_GPMU_GPMU_VOLTAGE,
|
|
(0x80000000 | device->pwrctrl.active_pwrlevel));
|
|
/* use the leakage to set this value at runtime */
|
|
kgsl_regwrite(device, A5XX_GPMU_BASE_LEAKAGE,
|
|
adreno_dev->lm_leakage);
|
|
|
|
/* Enable the power threshold and set it to 6000m */
|
|
kgsl_regwrite(device, A5XX_GPMU_GPMU_PWR_THRESHOLD,
|
|
0x80000000 | lm_limit(adreno_dev));
|
|
|
|
kgsl_regwrite(device, A5XX_GPMU_BEC_ENABLE, 0x10001FFF);
|
|
kgsl_regwrite(device, A5XX_GDPM_CONFIG1, 0x00201FF1);
|
|
|
|
/* Send an initial message to the GPMU with the LM voltage table */
|
|
kgsl_regwrite(device, AGC_MSG_STATE, 1);
|
|
kgsl_regwrite(device, AGC_MSG_COMMAND, AGC_POWER_CONFIG_PRODUCTION_ID);
|
|
length = _write_voltage_table(adreno_dev, AGC_MSG_PAYLOAD);
|
|
kgsl_regwrite(device, AGC_MSG_PAYLOAD_SIZE, length * sizeof(uint32_t));
|
|
kgsl_regwrite(device, AGC_INIT_MSG_MAGIC, AGC_INIT_MSG_VALUE);
|
|
}
|
|
|
|
/*
|
|
* a5xx_lm_enable() - Enable the LM/DPM feature on the GPMU
|
|
* @adreno_dev: The adreno device pointer
|
|
*/
|
|
static void a530_lm_enable(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_LM) ||
|
|
!test_bit(ADRENO_LM_CTRL, &adreno_dev->pwrctrl_flag))
|
|
return;
|
|
|
|
/* If no sequence properly initialized, return */
|
|
if (adreno_dev->lm_sequence == NULL)
|
|
return;
|
|
|
|
kgsl_regwrite(device, A5XX_GDPM_INT_MASK, 0x00000000);
|
|
kgsl_regwrite(device, A5XX_GDPM_INT_EN, 0x0000000A);
|
|
kgsl_regwrite(device, A5XX_GPMU_GPMU_VOLTAGE_INTR_EN_MASK, 0x00000001);
|
|
kgsl_regwrite(device, A5XX_GPMU_TEMP_THRESHOLD_INTR_EN_MASK,
|
|
0x00050000);
|
|
kgsl_regwrite(device, A5XX_GPMU_THROTTLE_UNMASK_FORCE_CTRL,
|
|
0x00030000);
|
|
|
|
if (adreno_is_a530(adreno_dev))
|
|
/* Program throttle control, do not enable idle DCS on v3+ */
|
|
kgsl_regwrite(device, A5XX_GPMU_CLOCK_THROTTLE_CTRL,
|
|
adreno_is_a530v2(adreno_dev) ? 0x00060011 : 0x00000011);
|
|
}
|
|
|
|
static void a540_lm_init(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
uint32_t agc_lm_config = AGC_BCL_DISABLED |
|
|
((ADRENO_CHIPID_PATCH(adreno_dev->chipid) & 0x3)
|
|
<< AGC_GPU_VERSION_SHIFT);
|
|
unsigned int r;
|
|
|
|
if (!test_bit(ADRENO_THROTTLING_CTRL, &adreno_dev->pwrctrl_flag))
|
|
agc_lm_config |= AGC_THROTTLE_DISABLE;
|
|
|
|
if (lm_on(adreno_dev)) {
|
|
agc_lm_config |=
|
|
AGC_LM_CONFIG_ENABLE_GPMU_ADAPTIVE |
|
|
AGC_LM_CONFIG_ISENSE_ENABLE;
|
|
|
|
kgsl_regread(device, A5XX_GPMU_TEMP_SENSOR_CONFIG, &r);
|
|
|
|
if ((r & GPMU_ISENSE_STATUS) == GPMU_ISENSE_END_POINT_CAL_ERR) {
|
|
KGSL_CORE_ERR(
|
|
"GPMU: ISENSE end point calibration failure\n");
|
|
agc_lm_config |= AGC_LM_CONFIG_ENABLE_ERROR;
|
|
}
|
|
}
|
|
|
|
kgsl_regwrite(device, AGC_MSG_STATE, 0x80000001);
|
|
kgsl_regwrite(device, AGC_MSG_COMMAND, AGC_POWER_CONFIG_PRODUCTION_ID);
|
|
(void) _write_voltage_table(adreno_dev, AGC_MSG_PAYLOAD);
|
|
kgsl_regwrite(device, AGC_MSG_PAYLOAD + AGC_LM_CONFIG, agc_lm_config);
|
|
kgsl_regwrite(device, AGC_MSG_PAYLOAD + AGC_LEVEL_CONFIG,
|
|
(unsigned int) ~(GENMASK(LM_DCVS_LIMIT, 0) |
|
|
GENMASK(16+LM_DCVS_LIMIT, 16)));
|
|
|
|
kgsl_regwrite(device, AGC_MSG_PAYLOAD_SIZE,
|
|
(AGC_LEVEL_CONFIG + 1) * sizeof(uint32_t));
|
|
kgsl_regwrite(device, AGC_INIT_MSG_MAGIC, AGC_INIT_MSG_VALUE);
|
|
|
|
kgsl_regwrite(device, A5XX_GPMU_GPMU_VOLTAGE,
|
|
(0x80000000 | device->pwrctrl.active_pwrlevel));
|
|
|
|
kgsl_regwrite(device, A5XX_GPMU_GPMU_PWR_THRESHOLD,
|
|
PWR_THRESHOLD_VALID | lm_limit(adreno_dev));
|
|
|
|
kgsl_regwrite(device, A5XX_GPMU_GPMU_VOLTAGE_INTR_EN_MASK,
|
|
VOLTAGE_INTR_EN);
|
|
}
|
|
|
|
|
|
static void a5xx_lm_enable(struct adreno_device *adreno_dev)
|
|
{
|
|
if (adreno_is_a530(adreno_dev))
|
|
a530_lm_enable(adreno_dev);
|
|
}
|
|
|
|
static void a5xx_lm_init(struct adreno_device *adreno_dev)
|
|
{
|
|
if (adreno_is_a530(adreno_dev))
|
|
a530_lm_init(adreno_dev);
|
|
else if (adreno_is_a540(adreno_dev))
|
|
a540_lm_init(adreno_dev);
|
|
}
|
|
|
|
static int gpmu_set_level(struct adreno_device *adreno_dev, unsigned int val)
|
|
{
|
|
unsigned int reg;
|
|
int retry = 100;
|
|
|
|
kgsl_regwrite(KGSL_DEVICE(adreno_dev), A5XX_GPMU_GPMU_VOLTAGE, val);
|
|
|
|
do {
|
|
kgsl_regread(KGSL_DEVICE(adreno_dev), A5XX_GPMU_GPMU_VOLTAGE,
|
|
®);
|
|
} while ((reg & 0x80000000) && retry--);
|
|
|
|
return (reg & 0x80000000) ? -ETIMEDOUT : 0;
|
|
}
|
|
|
|
/*
|
|
* a5xx_pwrlevel_change_settings() - Program the hardware during power level
|
|
* transitions
|
|
* @adreno_dev: The adreno device pointer
|
|
* @prelevel: The previous power level
|
|
* @postlevel: The new power level
|
|
* @post: True if called after the clock change has taken effect
|
|
*/
|
|
static void a5xx_pwrlevel_change_settings(struct adreno_device *adreno_dev,
|
|
unsigned int prelevel, unsigned int postlevel,
|
|
bool post)
|
|
{
|
|
int on = 0;
|
|
|
|
/*
|
|
* On pre A540 HW only call through if PPD or LMx
|
|
* is supported and enabled
|
|
*/
|
|
if (ADRENO_FEATURE(adreno_dev, ADRENO_PPD) &&
|
|
test_bit(ADRENO_PPD_CTRL, &adreno_dev->pwrctrl_flag))
|
|
on = ADRENO_PPD;
|
|
|
|
if (ADRENO_FEATURE(adreno_dev, ADRENO_LM) &&
|
|
test_bit(ADRENO_LM_CTRL, &adreno_dev->pwrctrl_flag))
|
|
on = ADRENO_LM;
|
|
|
|
/* On 540+ HW call through unconditionally as long as GPMU is enabled */
|
|
if (ADRENO_FEATURE(adreno_dev, ADRENO_GPMU)) {
|
|
if (adreno_is_a540(adreno_dev))
|
|
on = ADRENO_GPMU;
|
|
}
|
|
|
|
if (!on)
|
|
return;
|
|
|
|
if (post == 0) {
|
|
if (gpmu_set_level(adreno_dev, (0x80000010 | postlevel)))
|
|
KGSL_CORE_ERR(
|
|
"GPMU pre powerlevel did not stabilize\n");
|
|
} else {
|
|
if (gpmu_set_level(adreno_dev, (0x80000000 | postlevel)))
|
|
KGSL_CORE_ERR(
|
|
"GPMU post powerlevel did not stabilize\n");
|
|
}
|
|
}
|
|
|
|
static void a5xx_clk_set_options(struct adreno_device *adreno_dev,
|
|
const char *name, struct clk *clk, bool on)
|
|
{
|
|
|
|
if (!adreno_is_a540(adreno_dev) && !adreno_is_a512(adreno_dev) &&
|
|
!adreno_is_a508(adreno_dev))
|
|
return;
|
|
|
|
/* Handle clock settings for GFX PSCBCs */
|
|
if (on) {
|
|
if (!strcmp(name, "mem_iface_clk")) {
|
|
clk_set_flags(clk, CLKFLAG_NORETAIN_PERIPH);
|
|
clk_set_flags(clk, CLKFLAG_NORETAIN_MEM);
|
|
} else if (!strcmp(name, "core_clk")) {
|
|
clk_set_flags(clk, CLKFLAG_RETAIN_PERIPH);
|
|
clk_set_flags(clk, CLKFLAG_RETAIN_MEM);
|
|
}
|
|
} else {
|
|
if (!strcmp(name, "core_clk")) {
|
|
clk_set_flags(clk, CLKFLAG_NORETAIN_PERIPH);
|
|
clk_set_flags(clk, CLKFLAG_NORETAIN_MEM);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void a5xx_count_throttles(struct adreno_device *adreno_dev,
|
|
uint64_t adj)
|
|
{
|
|
if (adreno_is_a530(adreno_dev))
|
|
kgsl_regread(KGSL_DEVICE(adreno_dev),
|
|
adreno_dev->lm_threshold_count,
|
|
&adreno_dev->lm_threshold_cross);
|
|
else if (adreno_is_a540(adreno_dev))
|
|
adreno_dev->lm_threshold_cross = adj;
|
|
}
|
|
|
|
static int a5xx_enable_pwr_counters(struct adreno_device *adreno_dev,
|
|
unsigned int counter)
|
|
{
|
|
/*
|
|
* On 5XX we have to emulate the PWR counters which are physically
|
|
* missing. Program countable 6 on RBBM_PERFCTR_RBBM_0 as a substitute
|
|
* for PWR:1. Don't emulate PWR:0 as nobody uses it and we don't want
|
|
* to take away too many of the generic RBBM counters.
|
|
*/
|
|
|
|
if (counter == 0)
|
|
return -EINVAL;
|
|
|
|
kgsl_regwrite(KGSL_DEVICE(adreno_dev), A5XX_RBBM_PERFCTR_RBBM_SEL_0, 6);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* FW driven idle 10% throttle */
|
|
#define IDLE_10PCT 0
|
|
/* number of cycles when clock is throttled by 50% (CRC) */
|
|
#define CRC_50PCT 1
|
|
/* number of cycles when clock is throttled by more than 50% (CRC) */
|
|
#define CRC_MORE50PCT 2
|
|
/* number of cycles when clock is throttle by less than 50% (CRC) */
|
|
#define CRC_LESS50PCT 3
|
|
|
|
static int64_t a5xx_read_throttling_counters(struct adreno_device *adreno_dev)
|
|
{
|
|
int i;
|
|
int64_t adj;
|
|
uint32_t th[ADRENO_GPMU_THROTTLE_COUNTERS];
|
|
struct adreno_busy_data *busy = &adreno_dev->busy_data;
|
|
|
|
if (!adreno_is_a540(adreno_dev))
|
|
return 0;
|
|
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_GPMU))
|
|
return 0;
|
|
|
|
if (!test_bit(ADRENO_THROTTLING_CTRL, &adreno_dev->pwrctrl_flag))
|
|
return 0;
|
|
|
|
for (i = 0; i < ADRENO_GPMU_THROTTLE_COUNTERS; i++) {
|
|
if (!adreno_dev->gpmu_throttle_counters[i])
|
|
return 0;
|
|
|
|
th[i] = counter_delta(KGSL_DEVICE(adreno_dev),
|
|
adreno_dev->gpmu_throttle_counters[i],
|
|
&busy->throttle_cycles[i]);
|
|
}
|
|
adj = th[CRC_MORE50PCT] - th[IDLE_10PCT];
|
|
adj = th[CRC_50PCT] + th[CRC_LESS50PCT] / 3 + (adj < 0 ? 0 : adj) * 3;
|
|
|
|
trace_kgsl_clock_throttling(
|
|
th[IDLE_10PCT], th[CRC_50PCT],
|
|
th[CRC_MORE50PCT], th[CRC_LESS50PCT],
|
|
adj);
|
|
return adj;
|
|
}
|
|
|
|
static void a5xx_enable_64bit(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
kgsl_regwrite(device, A5XX_CP_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_VSC_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_GRAS_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_RB_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_PC_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_HLSQ_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_VFD_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_VPC_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_UCHE_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_SP_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_TPL1_ADDR_MODE_CNTL, 0x1);
|
|
kgsl_regwrite(device, A5XX_RBBM_SECVID_TSB_ADDR_MODE_CNTL, 0x1);
|
|
}
|
|
|
|
/*
|
|
* a5xx_gpmu_reset() - Re-enable GPMU based power features and restart GPMU
|
|
* @work: Pointer to the work struct for gpmu reset
|
|
*
|
|
* Load the GPMU microcode, set up any features such as hardware clock gating
|
|
* or IFPC, and take the GPMU out of reset.
|
|
*/
|
|
static void a5xx_gpmu_reset(struct work_struct *work)
|
|
{
|
|
struct adreno_device *adreno_dev = container_of(work,
|
|
struct adreno_device, gpmu_work);
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
if (test_bit(ADRENO_DEVICE_GPMU_INITIALIZED, &adreno_dev->priv))
|
|
return;
|
|
|
|
/*
|
|
* If GPMU has already experienced a restart or is in the process of it
|
|
* after the watchdog timeout, then there is no need to reset GPMU
|
|
* again.
|
|
*/
|
|
if (device->state != KGSL_STATE_NAP &&
|
|
device->state != KGSL_STATE_AWARE &&
|
|
device->state != KGSL_STATE_ACTIVE)
|
|
return;
|
|
|
|
mutex_lock(&device->mutex);
|
|
|
|
if (device->state == KGSL_STATE_NAP)
|
|
kgsl_pwrctrl_change_state(device, KGSL_STATE_AWARE);
|
|
|
|
if (a5xx_regulator_enable(adreno_dev))
|
|
goto out;
|
|
|
|
/* Soft reset of the GPMU block */
|
|
kgsl_regwrite(device, A5XX_RBBM_BLOCK_SW_RESET_CMD, BIT(16));
|
|
|
|
/* GPU comes up in secured mode, make it unsecured by default */
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_CONTENT_PROTECTION))
|
|
kgsl_regwrite(device, A5XX_RBBM_SECVID_TRUST_CNTL, 0x0);
|
|
|
|
|
|
a5xx_gpmu_init(adreno_dev);
|
|
|
|
out:
|
|
mutex_unlock(&device->mutex);
|
|
}
|
|
|
|
static void _setup_throttling_counters(struct adreno_device *adreno_dev)
|
|
{
|
|
int i, ret;
|
|
|
|
if (!adreno_is_a540(adreno_dev))
|
|
return;
|
|
|
|
if (!ADRENO_FEATURE(adreno_dev, ADRENO_GPMU))
|
|
return;
|
|
|
|
for (i = 0; i < ADRENO_GPMU_THROTTLE_COUNTERS; i++) {
|
|
/* reset throttled cycles ivalue */
|
|
adreno_dev->busy_data.throttle_cycles[i] = 0;
|
|
|
|
if (adreno_dev->gpmu_throttle_counters[i] != 0)
|
|
continue;
|
|
ret = adreno_perfcounter_get(adreno_dev,
|
|
KGSL_PERFCOUNTER_GROUP_GPMU_PWR,
|
|
ADRENO_GPMU_THROTTLE_COUNTERS_BASE_REG + i,
|
|
&adreno_dev->gpmu_throttle_counters[i],
|
|
NULL,
|
|
PERFCOUNTER_FLAG_KERNEL);
|
|
WARN_ONCE(ret, "Unable to get clock throttling counter %x\n",
|
|
ADRENO_GPMU_THROTTLE_COUNTERS_BASE_REG + i);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* a5xx_start() - Device start
|
|
* @adreno_dev: Pointer to adreno device
|
|
*
|
|
* a5xx device start
|
|
*/
|
|
static void a5xx_start(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
struct adreno_gpudev *gpudev = ADRENO_GPU_DEVICE(adreno_dev);
|
|
unsigned int bit;
|
|
int ret;
|
|
|
|
if (adreno_is_a530(adreno_dev) && ADRENO_FEATURE(adreno_dev, ADRENO_LM)
|
|
&& adreno_dev->lm_threshold_count == 0) {
|
|
|
|
ret = adreno_perfcounter_get(adreno_dev,
|
|
KGSL_PERFCOUNTER_GROUP_GPMU_PWR, 27,
|
|
&adreno_dev->lm_threshold_count, NULL,
|
|
PERFCOUNTER_FLAG_KERNEL);
|
|
/* Ignore noncritical ret - used for debugfs */
|
|
if (ret)
|
|
adreno_dev->lm_threshold_count = 0;
|
|
}
|
|
|
|
_setup_throttling_counters(adreno_dev);
|
|
|
|
adreno_vbif_start(adreno_dev, a5xx_vbif_platforms,
|
|
ARRAY_SIZE(a5xx_vbif_platforms));
|
|
|
|
/* Make all blocks contribute to the GPU BUSY perf counter */
|
|
kgsl_regwrite(device, A5XX_RBBM_PERFCTR_GPU_BUSY_MASKED, 0xFFFFFFFF);
|
|
|
|
/*
|
|
* Enable the RBBM error reporting bits. This lets us get
|
|
* useful information on failure
|
|
*/
|
|
kgsl_regwrite(device, A5XX_RBBM_AHB_CNTL0, 0x00000001);
|
|
|
|
if (ADRENO_QUIRK(adreno_dev, ADRENO_QUIRK_FAULT_DETECT_MASK)) {
|
|
/*
|
|
* We have 4 RB units, and only RB0 activity signals are
|
|
* working correctly. Mask out RB1-3 activity signals
|
|
* from the HW hang detection logic as per
|
|
* recommendation of hardware team.
|
|
*/
|
|
kgsl_regwrite(device, A5XX_RBBM_INTERFACE_HANG_MASK_CNTL11,
|
|
0xF0000000);
|
|
kgsl_regwrite(device, A5XX_RBBM_INTERFACE_HANG_MASK_CNTL12,
|
|
0xFFFFFFFF);
|
|
kgsl_regwrite(device, A5XX_RBBM_INTERFACE_HANG_MASK_CNTL13,
|
|
0xFFFFFFFF);
|
|
kgsl_regwrite(device, A5XX_RBBM_INTERFACE_HANG_MASK_CNTL14,
|
|
0xFFFFFFFF);
|
|
kgsl_regwrite(device, A5XX_RBBM_INTERFACE_HANG_MASK_CNTL15,
|
|
0xFFFFFFFF);
|
|
kgsl_regwrite(device, A5XX_RBBM_INTERFACE_HANG_MASK_CNTL16,
|
|
0xFFFFFFFF);
|
|
kgsl_regwrite(device, A5XX_RBBM_INTERFACE_HANG_MASK_CNTL17,
|
|
0xFFFFFFFF);
|
|
kgsl_regwrite(device, A5XX_RBBM_INTERFACE_HANG_MASK_CNTL18,
|
|
0xFFFFFFFF);
|
|
}
|
|
|
|
/*
|
|
* Turn on hang detection for a530 v2 and beyond. This spews a
|
|
* lot of useful information into the RBBM registers on a hang.
|
|
*/
|
|
if (!adreno_is_a530v1(adreno_dev)) {
|
|
|
|
set_bit(ADRENO_DEVICE_HANG_INTR, &adreno_dev->priv);
|
|
gpudev->irq->mask |= (1 << A5XX_INT_MISC_HANG_DETECT);
|
|
/*
|
|
* Set hang detection threshold to 4 million cycles
|
|
* (0x3FFFF*16)
|
|
*/
|
|
kgsl_regwrite(device, A5XX_RBBM_INTERFACE_HANG_INT_CNTL,
|
|
(1 << 30) | 0x3FFFF);
|
|
}
|
|
|
|
|
|
/* Turn on performance counters */
|
|
kgsl_regwrite(device, A5XX_RBBM_PERFCTR_CNTL, 0x01);
|
|
|
|
/*
|
|
* This is to increase performance by restricting VFD's cache access,
|
|
* so that LRZ and other data get evicted less.
|
|
*/
|
|
kgsl_regwrite(device, A5XX_UCHE_CACHE_WAYS, 0x02);
|
|
|
|
/*
|
|
* Set UCHE_WRITE_THRU_BASE to the UCHE_TRAP_BASE effectively
|
|
* disabling L2 bypass
|
|
*/
|
|
kgsl_regwrite(device, A5XX_UCHE_TRAP_BASE_LO, 0xffff0000);
|
|
kgsl_regwrite(device, A5XX_UCHE_TRAP_BASE_HI, 0x0001ffff);
|
|
kgsl_regwrite(device, A5XX_UCHE_WRITE_THRU_BASE_LO, 0xffff0000);
|
|
kgsl_regwrite(device, A5XX_UCHE_WRITE_THRU_BASE_HI, 0x0001ffff);
|
|
|
|
/* Program the GMEM VA range for the UCHE path */
|
|
kgsl_regwrite(device, A5XX_UCHE_GMEM_RANGE_MIN_LO,
|
|
ADRENO_UCHE_GMEM_BASE);
|
|
kgsl_regwrite(device, A5XX_UCHE_GMEM_RANGE_MIN_HI, 0x0);
|
|
kgsl_regwrite(device, A5XX_UCHE_GMEM_RANGE_MAX_LO,
|
|
ADRENO_UCHE_GMEM_BASE +
|
|
adreno_dev->gmem_size - 1);
|
|
kgsl_regwrite(device, A5XX_UCHE_GMEM_RANGE_MAX_HI, 0x0);
|
|
|
|
/*
|
|
* Below CP registers are 0x0 by default, program init
|
|
* values based on a5xx flavor.
|
|
*/
|
|
if (adreno_is_a504_to_a506(adreno_dev) || adreno_is_a508(adreno_dev)) {
|
|
kgsl_regwrite(device, A5XX_CP_MEQ_THRESHOLDS, 0x20);
|
|
kgsl_regwrite(device, A5XX_CP_MERCIU_SIZE, 0x400);
|
|
kgsl_regwrite(device, A5XX_CP_ROQ_THRESHOLDS_2, 0x40000030);
|
|
kgsl_regwrite(device, A5XX_CP_ROQ_THRESHOLDS_1, 0x20100D0A);
|
|
} else if (adreno_is_a510(adreno_dev)) {
|
|
kgsl_regwrite(device, A5XX_CP_MEQ_THRESHOLDS, 0x20);
|
|
kgsl_regwrite(device, A5XX_CP_MERCIU_SIZE, 0x20);
|
|
kgsl_regwrite(device, A5XX_CP_ROQ_THRESHOLDS_2, 0x40000030);
|
|
kgsl_regwrite(device, A5XX_CP_ROQ_THRESHOLDS_1, 0x20100D0A);
|
|
} else if (adreno_is_a540(adreno_dev) || adreno_is_a512(adreno_dev)) {
|
|
kgsl_regwrite(device, A5XX_CP_MEQ_THRESHOLDS, 0x40);
|
|
kgsl_regwrite(device, A5XX_CP_MERCIU_SIZE, 0x400);
|
|
kgsl_regwrite(device, A5XX_CP_ROQ_THRESHOLDS_2, 0x80000060);
|
|
kgsl_regwrite(device, A5XX_CP_ROQ_THRESHOLDS_1, 0x40201B16);
|
|
} else {
|
|
kgsl_regwrite(device, A5XX_CP_MEQ_THRESHOLDS, 0x40);
|
|
kgsl_regwrite(device, A5XX_CP_MERCIU_SIZE, 0x40);
|
|
kgsl_regwrite(device, A5XX_CP_ROQ_THRESHOLDS_2, 0x80000060);
|
|
kgsl_regwrite(device, A5XX_CP_ROQ_THRESHOLDS_1, 0x40201B16);
|
|
}
|
|
|
|
/*
|
|
* vtxFifo and primFifo thresholds default values
|
|
* are different.
|
|
*/
|
|
if (adreno_is_a504_to_a506(adreno_dev) || adreno_is_a508(adreno_dev))
|
|
kgsl_regwrite(device, A5XX_PC_DBG_ECO_CNTL,
|
|
(0x100 << 11 | 0x100 << 22));
|
|
else if (adreno_is_a510(adreno_dev) || adreno_is_a512(adreno_dev))
|
|
kgsl_regwrite(device, A5XX_PC_DBG_ECO_CNTL,
|
|
(0x200 << 11 | 0x200 << 22));
|
|
else
|
|
kgsl_regwrite(device, A5XX_PC_DBG_ECO_CNTL,
|
|
(0x400 << 11 | 0x300 << 22));
|
|
|
|
/*
|
|
* A5x USP LDST non valid pixel wrongly update read combine offset
|
|
* In A5xx we added optimization for read combine. There could be cases
|
|
* on a530 v1 there is no valid pixel but the active masks is not
|
|
* cleared and the offset can be wrongly updated if the invalid address
|
|
* can be combined. The wrongly latched value will make the returning
|
|
* data got shifted at wrong offset. workaround this issue by disabling
|
|
* LD combine, bit[25] of SP_DBG_ECO_CNTL (sp chicken bit[17]) need to
|
|
* be set to 1, default is 0(enable)
|
|
*/
|
|
if (adreno_is_a530v1(adreno_dev))
|
|
kgsl_regrmw(device, A5XX_SP_DBG_ECO_CNTL, 0, (1 << 25));
|
|
|
|
if (ADRENO_QUIRK(adreno_dev, ADRENO_QUIRK_TWO_PASS_USE_WFI)) {
|
|
/*
|
|
* Set TWOPASSUSEWFI in A5XX_PC_DBG_ECO_CNTL for
|
|
* microcodes after v77
|
|
*/
|
|
if ((adreno_compare_pfp_version(adreno_dev, 0x5FF077) >= 0))
|
|
kgsl_regrmw(device, A5XX_PC_DBG_ECO_CNTL, 0, (1 << 8));
|
|
}
|
|
|
|
if (ADRENO_QUIRK(adreno_dev, ADRENO_QUIRK_DISABLE_RB_DP2CLOCKGATING)) {
|
|
/*
|
|
* Disable RB sampler datapath DP2 clock gating
|
|
* optimization for 1-SP GPU's, by default it is enabled.
|
|
*/
|
|
kgsl_regrmw(device, A5XX_RB_DBG_ECO_CNT, 0, (1 << 9));
|
|
}
|
|
/*
|
|
* Disable UCHE global filter as SP can invalidate/flush
|
|
* independently
|
|
*/
|
|
kgsl_regwrite(device, A5XX_UCHE_MODE_CNTL, BIT(29));
|
|
/* Set the USE_RETENTION_FLOPS chicken bit */
|
|
kgsl_regwrite(device, A5XX_CP_CHICKEN_DBG, 0x02000000);
|
|
|
|
/* Enable ISDB mode if requested */
|
|
if (test_bit(ADRENO_DEVICE_ISDB_ENABLED, &adreno_dev->priv)) {
|
|
if (!kgsl_active_count_get(device)) {
|
|
/*
|
|
* Disable ME/PFP split timeouts when the debugger is
|
|
* enabled because the CP doesn't know when a shader is
|
|
* in active debug
|
|
*/
|
|
kgsl_regwrite(device, A5XX_RBBM_AHB_CNTL1, 0x06FFFFFF);
|
|
|
|
/* Force the SP0/SP1 clocks on to enable ISDB */
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL_SP0, 0x0);
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL_SP1, 0x0);
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL_SP2, 0x0);
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL_SP3, 0x0);
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL2_SP0, 0x0);
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL2_SP1, 0x0);
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL2_SP2, 0x0);
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL2_SP3, 0x0);
|
|
|
|
/* disable HWCG */
|
|
kgsl_regwrite(device, A5XX_RBBM_CLOCK_CNTL, 0x0);
|
|
kgsl_regwrite(device, A5XX_RBBM_ISDB_CNT, 0x0);
|
|
} else
|
|
KGSL_CORE_ERR(
|
|
"Active count failed while turning on ISDB.");
|
|
} else {
|
|
/* if not in ISDB mode enable ME/PFP split notification */
|
|
kgsl_regwrite(device, A5XX_RBBM_AHB_CNTL1, 0xA6FFFFFF);
|
|
}
|
|
|
|
kgsl_regwrite(device, A5XX_RBBM_AHB_CNTL2, 0x0000003F);
|
|
|
|
if (!of_property_read_u32(device->pdev->dev.of_node,
|
|
"qcom,highest-bank-bit", &bit)) {
|
|
if (bit >= 13 && bit <= 16) {
|
|
bit = (bit - 13) & 0x03;
|
|
|
|
/*
|
|
* Program the highest DDR bank bit that was passed in
|
|
* from the DT in a handful of registers. Some of these
|
|
* registers will also be written by the UMD, but we
|
|
* want to program them in case we happen to use the
|
|
* UCHE before the UMD does
|
|
*/
|
|
|
|
kgsl_regwrite(device, A5XX_TPL1_MODE_CNTL, bit << 7);
|
|
kgsl_regwrite(device, A5XX_RB_MODE_CNTL, bit << 1);
|
|
if (adreno_is_a540(adreno_dev) ||
|
|
adreno_is_a512(adreno_dev))
|
|
kgsl_regwrite(device, A5XX_UCHE_DBG_ECO_CNTL_2,
|
|
bit);
|
|
}
|
|
|
|
}
|
|
|
|
/* Disable All flat shading optimization */
|
|
kgsl_regrmw(device, A5XX_VPC_DBG_ECO_CNTL, 0, 0x1 << 10);
|
|
|
|
/*
|
|
* VPC corner case with local memory load kill leads to corrupt
|
|
* internal state. Normal Disable does not work for all a5x chips.
|
|
* So do the following setting to disable it.
|
|
*/
|
|
if (ADRENO_QUIRK(adreno_dev, ADRENO_QUIRK_DISABLE_LMLOADKILL)) {
|
|
kgsl_regrmw(device, A5XX_VPC_DBG_ECO_CNTL, 0, 0x1 << 23);
|
|
kgsl_regrmw(device, A5XX_HLSQ_DBG_ECO_CNTL, 0x1 << 18, 0);
|
|
}
|
|
|
|
a5xx_preemption_start(adreno_dev);
|
|
a5xx_protect_init(adreno_dev);
|
|
}
|
|
|
|
/*
|
|
* Follow the ME_INIT sequence with a preemption yield to allow the GPU to move
|
|
* to a different ringbuffer, if desired
|
|
*/
|
|
static int _preemption_init(
|
|
struct adreno_device *adreno_dev,
|
|
struct adreno_ringbuffer *rb, unsigned int *cmds,
|
|
struct kgsl_context *context)
|
|
{
|
|
unsigned int *cmds_orig = cmds;
|
|
uint64_t gpuaddr = rb->preemption_desc.gpuaddr;
|
|
|
|
/* Turn CP protection OFF */
|
|
*cmds++ = cp_type7_packet(CP_SET_PROTECTED_MODE, 1);
|
|
*cmds++ = 0;
|
|
/*
|
|
* CP during context switch will save context switch info to
|
|
* a5xx_cp_preemption_record pointed by CONTEXT_SWITCH_SAVE_ADDR
|
|
*/
|
|
*cmds++ = cp_type4_packet(A5XX_CP_CONTEXT_SWITCH_SAVE_ADDR_LO, 1);
|
|
*cmds++ = lower_32_bits(gpuaddr);
|
|
*cmds++ = cp_type4_packet(A5XX_CP_CONTEXT_SWITCH_SAVE_ADDR_HI, 1);
|
|
*cmds++ = upper_32_bits(gpuaddr);
|
|
|
|
/* Turn CP protection ON */
|
|
*cmds++ = cp_type7_packet(CP_SET_PROTECTED_MODE, 1);
|
|
*cmds++ = 1;
|
|
|
|
*cmds++ = cp_type7_packet(CP_PREEMPT_ENABLE_GLOBAL, 1);
|
|
*cmds++ = 0;
|
|
|
|
*cmds++ = cp_type7_packet(CP_PREEMPT_ENABLE_LOCAL, 1);
|
|
*cmds++ = 1;
|
|
|
|
/* Enable yield in RB only */
|
|
*cmds++ = cp_type7_packet(CP_YIELD_ENABLE, 1);
|
|
*cmds++ = 1;
|
|
|
|
*cmds++ = cp_type7_packet(CP_CONTEXT_SWITCH_YIELD, 4);
|
|
cmds += cp_gpuaddr(adreno_dev, cmds, 0x0);
|
|
*cmds++ = 0;
|
|
/* generate interrupt on preemption completion */
|
|
*cmds++ = 1;
|
|
|
|
return cmds - cmds_orig;
|
|
}
|
|
|
|
static int a5xx_post_start(struct adreno_device *adreno_dev)
|
|
{
|
|
int ret;
|
|
unsigned int *cmds, *start;
|
|
struct adreno_ringbuffer *rb = adreno_dev->cur_rb;
|
|
|
|
if (!adreno_is_a530(adreno_dev) &&
|
|
!adreno_is_preemption_enabled(adreno_dev))
|
|
return 0;
|
|
|
|
cmds = adreno_ringbuffer_allocspace(rb, 42);
|
|
if (IS_ERR(cmds)) {
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
KGSL_DRV_ERR(device, "error allocating preemtion init cmds");
|
|
return PTR_ERR(cmds);
|
|
}
|
|
start = cmds;
|
|
|
|
/*
|
|
* Send a pipeline stat event whenever the GPU gets powered up
|
|
* to cause misbehaving perf counters to start ticking
|
|
*/
|
|
if (adreno_is_a530(adreno_dev)) {
|
|
*cmds++ = cp_packet(adreno_dev, CP_EVENT_WRITE, 1);
|
|
*cmds++ = 0xF;
|
|
}
|
|
|
|
if (adreno_is_preemption_enabled(adreno_dev)) {
|
|
cmds += _preemption_init(adreno_dev, rb, cmds, NULL);
|
|
rb->_wptr = rb->_wptr - (42 - (cmds - start));
|
|
ret = adreno_ringbuffer_submit_spin_nosync(rb, NULL, 2000);
|
|
} else {
|
|
rb->_wptr = rb->_wptr - (42 - (cmds - start));
|
|
ret = adreno_ringbuffer_submit_spin(rb, NULL, 2000);
|
|
}
|
|
|
|
if (ret)
|
|
adreno_spin_idle_debug(adreno_dev,
|
|
"hw initialization failed to idle\n");
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int a5xx_gpmu_init(struct adreno_device *adreno_dev)
|
|
{
|
|
int ret;
|
|
|
|
/* Set up LM before initializing the GPMU */
|
|
a5xx_lm_init(adreno_dev);
|
|
|
|
/* Enable SPTP based power collapse before enabling GPMU */
|
|
a5xx_enable_pc(adreno_dev);
|
|
|
|
ret = a5xx_gpmu_start(adreno_dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* Enable limits management */
|
|
a5xx_lm_enable(adreno_dev);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* a5xx_microcode_load() - Load microcode
|
|
* @adreno_dev: Pointer to adreno device
|
|
*/
|
|
static int a5xx_microcode_load(struct adreno_device *adreno_dev)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
struct adreno_firmware *pm4_fw = ADRENO_FW(adreno_dev, ADRENO_FW_PM4);
|
|
struct adreno_firmware *pfp_fw = ADRENO_FW(adreno_dev, ADRENO_FW_PFP);
|
|
uint64_t gpuaddr;
|
|
int ret = 0;
|
|
|
|
gpuaddr = pm4_fw->memdesc.gpuaddr;
|
|
kgsl_regwrite(device, A5XX_CP_PM4_INSTR_BASE_LO,
|
|
lower_32_bits(gpuaddr));
|
|
kgsl_regwrite(device, A5XX_CP_PM4_INSTR_BASE_HI,
|
|
upper_32_bits(gpuaddr));
|
|
|
|
gpuaddr = pfp_fw->memdesc.gpuaddr;
|
|
kgsl_regwrite(device, A5XX_CP_PFP_INSTR_BASE_LO,
|
|
lower_32_bits(gpuaddr));
|
|
kgsl_regwrite(device, A5XX_CP_PFP_INSTR_BASE_HI,
|
|
upper_32_bits(gpuaddr));
|
|
|
|
/*
|
|
* Do not invoke to load zap shader if MMU does
|
|
* not support secure mode.
|
|
*/
|
|
if (!device->mmu.secured)
|
|
return 0;
|
|
|
|
/*
|
|
* Resume call to write the zap shader base address into the
|
|
* appropriate register,
|
|
* skip if retention is supported for the CPZ register
|
|
*/
|
|
if (adreno_dev->zap_handle_ptr && !(ADRENO_FEATURE(adreno_dev,
|
|
ADRENO_CPZ_RETENTION))) {
|
|
struct scm_desc desc = {0};
|
|
|
|
desc.args[0] = 0;
|
|
desc.args[1] = 13;
|
|
desc.arginfo = SCM_ARGS(2);
|
|
|
|
ret = scm_call2(SCM_SIP_FNID(SCM_SVC_BOOT, 0xA), &desc);
|
|
if (ret) {
|
|
pr_err("SCM resume call failed with error %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
}
|
|
|
|
/* Load the zap shader firmware through PIL if its available */
|
|
if (adreno_dev->gpucore->zap_name && !adreno_dev->zap_handle_ptr) {
|
|
adreno_dev->zap_handle_ptr =
|
|
subsystem_get(adreno_dev->gpucore->zap_name);
|
|
|
|
/* Return error if the zap shader cannot be loaded */
|
|
if (IS_ERR_OR_NULL(adreno_dev->zap_handle_ptr)) {
|
|
ret = (adreno_dev->zap_handle_ptr == NULL) ?
|
|
-ENODEV : PTR_ERR(adreno_dev->zap_handle_ptr);
|
|
adreno_dev->zap_handle_ptr = NULL;
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void a5xx_zap_shader_unload(struct adreno_device *adreno_dev)
|
|
{
|
|
if (!IS_ERR_OR_NULL(adreno_dev->zap_handle_ptr)) {
|
|
subsystem_put(adreno_dev->zap_handle_ptr);
|
|
adreno_dev->zap_handle_ptr = NULL;
|
|
}
|
|
}
|
|
|
|
static int _me_init_ucode_workarounds(struct adreno_device *adreno_dev)
|
|
{
|
|
switch (ADRENO_GPUREV(adreno_dev)) {
|
|
case ADRENO_REV_A510:
|
|
return 0x00000001; /* Ucode workaround for token end syncs */
|
|
case ADRENO_REV_A504:
|
|
case ADRENO_REV_A505:
|
|
case ADRENO_REV_A506:
|
|
case ADRENO_REV_A530:
|
|
/*
|
|
* Ucode workarounds for token end syncs,
|
|
* WFI after every direct-render 3D mode draw and
|
|
* WFI after every 2D Mode 3 draw.
|
|
*/
|
|
return 0x0000000B;
|
|
case ADRENO_REV_A540:
|
|
/*
|
|
* WFI after every direct-render 3D mode draw and
|
|
* WFI after every 2D Mode 3 draw. This is needed
|
|
* only on a540v1.
|
|
*/
|
|
if (adreno_is_a540v1(adreno_dev))
|
|
return 0x0000000A;
|
|
default:
|
|
return 0x00000000; /* No ucode workarounds enabled */
|
|
}
|
|
}
|
|
|
|
/*
|
|
* CP_INIT_MAX_CONTEXT bit tells if the multiple hardware contexts can
|
|
* be used at once of if they should be serialized
|
|
*/
|
|
#define CP_INIT_MAX_CONTEXT BIT(0)
|
|
|
|
/* Enables register protection mode */
|
|
#define CP_INIT_ERROR_DETECTION_CONTROL BIT(1)
|
|
|
|
/* Header dump information */
|
|
#define CP_INIT_HEADER_DUMP BIT(2) /* Reserved */
|
|
|
|
/* Default Reset states enabled for PFP and ME */
|
|
#define CP_INIT_DEFAULT_RESET_STATE BIT(3)
|
|
|
|
/* Drawcall filter range */
|
|
#define CP_INIT_DRAWCALL_FILTER_RANGE BIT(4)
|
|
|
|
/* Ucode workaround masks */
|
|
#define CP_INIT_UCODE_WORKAROUND_MASK BIT(5)
|
|
|
|
#define CP_INIT_MASK (CP_INIT_MAX_CONTEXT | \
|
|
CP_INIT_ERROR_DETECTION_CONTROL | \
|
|
CP_INIT_HEADER_DUMP | \
|
|
CP_INIT_DEFAULT_RESET_STATE | \
|
|
CP_INIT_UCODE_WORKAROUND_MASK)
|
|
|
|
static void _set_ordinals(struct adreno_device *adreno_dev,
|
|
unsigned int *cmds, unsigned int count)
|
|
{
|
|
unsigned int *start = cmds;
|
|
|
|
/* Enabled ordinal mask */
|
|
*cmds++ = CP_INIT_MASK;
|
|
|
|
if (CP_INIT_MASK & CP_INIT_MAX_CONTEXT) {
|
|
/*
|
|
* Multiple HW ctxs are unreliable on a530v1,
|
|
* use single hw context.
|
|
* Use multiple contexts if bit set, otherwise serialize:
|
|
* 3D (bit 0) 2D (bit 1)
|
|
*/
|
|
if (adreno_is_a530v1(adreno_dev))
|
|
*cmds++ = 0x00000000;
|
|
else
|
|
*cmds++ = 0x00000003;
|
|
}
|
|
|
|
if (CP_INIT_MASK & CP_INIT_ERROR_DETECTION_CONTROL)
|
|
*cmds++ = 0x20000000;
|
|
|
|
if (CP_INIT_MASK & CP_INIT_HEADER_DUMP) {
|
|
/* Header dump address */
|
|
*cmds++ = 0x00000000;
|
|
/* Header dump enable and dump size */
|
|
*cmds++ = 0x00000000;
|
|
}
|
|
|
|
if (CP_INIT_MASK & CP_INIT_DRAWCALL_FILTER_RANGE) {
|
|
/* Start range */
|
|
*cmds++ = 0x00000000;
|
|
/* End range (inclusive) */
|
|
*cmds++ = 0x00000000;
|
|
}
|
|
|
|
if (CP_INIT_MASK & CP_INIT_UCODE_WORKAROUND_MASK)
|
|
*cmds++ = _me_init_ucode_workarounds(adreno_dev);
|
|
|
|
/* Pad rest of the cmds with 0's */
|
|
while ((unsigned int)(cmds - start) < count)
|
|
*cmds++ = 0x0;
|
|
}
|
|
|
|
int a5xx_critical_packet_submit(struct adreno_device *adreno_dev,
|
|
struct adreno_ringbuffer *rb)
|
|
{
|
|
unsigned int *cmds;
|
|
int ret;
|
|
|
|
if (!critical_packet_constructed)
|
|
return 0;
|
|
|
|
cmds = adreno_ringbuffer_allocspace(rb, 4);
|
|
if (IS_ERR(cmds))
|
|
return PTR_ERR(cmds);
|
|
|
|
*cmds++ = cp_mem_packet(adreno_dev, CP_INDIRECT_BUFFER_PFE, 2, 1);
|
|
cmds += cp_gpuaddr(adreno_dev, cmds, crit_pkts.gpuaddr);
|
|
*cmds++ = crit_pkts_dwords;
|
|
|
|
ret = adreno_ringbuffer_submit_spin(rb, NULL, 20);
|
|
if (ret)
|
|
adreno_spin_idle_debug(adreno_dev,
|
|
"Critical packet submission failed to idle\n");
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* a5xx_send_me_init() - Initialize ringbuffer
|
|
* @adreno_dev: Pointer to adreno device
|
|
* @rb: Pointer to the ringbuffer of device
|
|
*
|
|
* Submit commands for ME initialization,
|
|
*/
|
|
static int a5xx_send_me_init(struct adreno_device *adreno_dev,
|
|
struct adreno_ringbuffer *rb)
|
|
{
|
|
unsigned int *cmds;
|
|
int ret;
|
|
|
|
cmds = adreno_ringbuffer_allocspace(rb, 9);
|
|
if (IS_ERR(cmds))
|
|
return PTR_ERR(cmds);
|
|
if (cmds == NULL)
|
|
return -ENOSPC;
|
|
|
|
*cmds++ = cp_type7_packet(CP_ME_INIT, 8);
|
|
|
|
_set_ordinals(adreno_dev, cmds, 8);
|
|
|
|
ret = adreno_ringbuffer_submit_spin(rb, NULL, 2000);
|
|
if (ret)
|
|
adreno_spin_idle_debug(adreno_dev,
|
|
"CP initialization failed to idle\n");
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* a5xx_rb_start() - Start the ringbuffer
|
|
* @adreno_dev: Pointer to adreno device
|
|
* @start_type: Warm or cold start
|
|
*/
|
|
static int a5xx_rb_start(struct adreno_device *adreno_dev,
|
|
unsigned int start_type)
|
|
{
|
|
struct adreno_ringbuffer *rb = ADRENO_CURRENT_RINGBUFFER(adreno_dev);
|
|
struct kgsl_device *device = &adreno_dev->dev;
|
|
uint64_t addr;
|
|
int ret;
|
|
|
|
addr = SCRATCH_RPTR_GPU_ADDR(device, rb->id);
|
|
|
|
adreno_writereg64(adreno_dev, ADRENO_REG_CP_RB_RPTR_ADDR_LO,
|
|
ADRENO_REG_CP_RB_RPTR_ADDR_HI, addr);
|
|
|
|
/*
|
|
* The size of the ringbuffer in the hardware is the log2
|
|
* representation of the size in quadwords (sizedwords / 2).
|
|
* Also disable the host RPTR shadow register as it might be unreliable
|
|
* in certain circumstances.
|
|
*/
|
|
|
|
adreno_writereg(adreno_dev, ADRENO_REG_CP_RB_CNTL,
|
|
A5XX_CP_RB_CNTL_DEFAULT);
|
|
|
|
adreno_writereg64(adreno_dev, ADRENO_REG_CP_RB_BASE,
|
|
ADRENO_REG_CP_RB_BASE_HI, rb->buffer_desc.gpuaddr);
|
|
|
|
ret = a5xx_microcode_load(adreno_dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* clear ME_HALT to start micro engine */
|
|
adreno_writereg(adreno_dev, ADRENO_REG_CP_ME_CNTL, 0);
|
|
|
|
ret = a5xx_send_me_init(adreno_dev, rb);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* GPU comes up in secured mode, make it unsecured by default */
|
|
ret = adreno_set_unsecured_mode(adreno_dev, rb);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = a5xx_gpmu_init(adreno_dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
a5xx_post_start(adreno_dev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int _load_firmware(struct kgsl_device *device, const char *fwfile,
|
|
struct adreno_firmware *firmware)
|
|
{
|
|
const struct firmware *fw = NULL;
|
|
int ret;
|
|
|
|
ret = request_firmware(&fw, fwfile, device->dev);
|
|
|
|
if (ret) {
|
|
KGSL_DRV_ERR(device, "request_firmware(%s) failed: %d\n",
|
|
fwfile, ret);
|
|
return ret;
|
|
}
|
|
|
|
ret = kgsl_allocate_global(device, &firmware->memdesc, fw->size - 4,
|
|
KGSL_MEMFLAGS_GPUREADONLY, 0, "ucode");
|
|
|
|
if (ret)
|
|
goto done;
|
|
|
|
memcpy(firmware->memdesc.hostptr, &fw->data[4], fw->size - 4);
|
|
firmware->size = (fw->size - 4) / sizeof(uint32_t);
|
|
firmware->version = adreno_get_ucode_version((u32 *)fw->data);
|
|
|
|
done:
|
|
release_firmware(fw);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* a5xx_microcode_read() - Read microcode
|
|
* @adreno_dev: Pointer to adreno device
|
|
*/
|
|
static int a5xx_microcode_read(struct adreno_device *adreno_dev)
|
|
{
|
|
int ret;
|
|
struct adreno_firmware *pm4_fw = ADRENO_FW(adreno_dev, ADRENO_FW_PM4);
|
|
struct adreno_firmware *pfp_fw = ADRENO_FW(adreno_dev, ADRENO_FW_PFP);
|
|
|
|
if (pm4_fw->memdesc.hostptr == NULL) {
|
|
ret = _load_firmware(KGSL_DEVICE(adreno_dev),
|
|
adreno_dev->gpucore->pm4fw_name, pm4_fw);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
if (pfp_fw->memdesc.hostptr == NULL) {
|
|
ret = _load_firmware(KGSL_DEVICE(adreno_dev),
|
|
adreno_dev->gpucore->pfpfw_name, pfp_fw);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
ret = _load_gpmu_firmware(adreno_dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
_load_regfile(adreno_dev);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_cp[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CP_0_LO,
|
|
A5XX_RBBM_PERFCTR_CP_0_HI, 0, A5XX_CP_PERFCTR_CP_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CP_1_LO,
|
|
A5XX_RBBM_PERFCTR_CP_1_HI, 1, A5XX_CP_PERFCTR_CP_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CP_2_LO,
|
|
A5XX_RBBM_PERFCTR_CP_2_HI, 2, A5XX_CP_PERFCTR_CP_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CP_3_LO,
|
|
A5XX_RBBM_PERFCTR_CP_3_HI, 3, A5XX_CP_PERFCTR_CP_SEL_3 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CP_4_LO,
|
|
A5XX_RBBM_PERFCTR_CP_4_HI, 4, A5XX_CP_PERFCTR_CP_SEL_4 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CP_5_LO,
|
|
A5XX_RBBM_PERFCTR_CP_5_HI, 5, A5XX_CP_PERFCTR_CP_SEL_5 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CP_6_LO,
|
|
A5XX_RBBM_PERFCTR_CP_6_HI, 6, A5XX_CP_PERFCTR_CP_SEL_6 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CP_7_LO,
|
|
A5XX_RBBM_PERFCTR_CP_7_HI, 7, A5XX_CP_PERFCTR_CP_SEL_7 },
|
|
};
|
|
|
|
/*
|
|
* Note that PERFCTR_RBBM_0 is missing - it is used to emulate the PWR counters.
|
|
* See below.
|
|
*/
|
|
static struct adreno_perfcount_register a5xx_perfcounters_rbbm[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RBBM_1_LO,
|
|
A5XX_RBBM_PERFCTR_RBBM_1_HI, 9, A5XX_RBBM_PERFCTR_RBBM_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RBBM_2_LO,
|
|
A5XX_RBBM_PERFCTR_RBBM_2_HI, 10, A5XX_RBBM_PERFCTR_RBBM_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RBBM_3_LO,
|
|
A5XX_RBBM_PERFCTR_RBBM_3_HI, 11, A5XX_RBBM_PERFCTR_RBBM_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_pc[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_PC_0_LO,
|
|
A5XX_RBBM_PERFCTR_PC_0_HI, 12, A5XX_PC_PERFCTR_PC_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_PC_1_LO,
|
|
A5XX_RBBM_PERFCTR_PC_1_HI, 13, A5XX_PC_PERFCTR_PC_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_PC_2_LO,
|
|
A5XX_RBBM_PERFCTR_PC_2_HI, 14, A5XX_PC_PERFCTR_PC_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_PC_3_LO,
|
|
A5XX_RBBM_PERFCTR_PC_3_HI, 15, A5XX_PC_PERFCTR_PC_SEL_3 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_PC_4_LO,
|
|
A5XX_RBBM_PERFCTR_PC_4_HI, 16, A5XX_PC_PERFCTR_PC_SEL_4 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_PC_5_LO,
|
|
A5XX_RBBM_PERFCTR_PC_5_HI, 17, A5XX_PC_PERFCTR_PC_SEL_5 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_PC_6_LO,
|
|
A5XX_RBBM_PERFCTR_PC_6_HI, 18, A5XX_PC_PERFCTR_PC_SEL_6 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_PC_7_LO,
|
|
A5XX_RBBM_PERFCTR_PC_7_HI, 19, A5XX_PC_PERFCTR_PC_SEL_7 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_vfd[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VFD_0_LO,
|
|
A5XX_RBBM_PERFCTR_VFD_0_HI, 20, A5XX_VFD_PERFCTR_VFD_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VFD_1_LO,
|
|
A5XX_RBBM_PERFCTR_VFD_1_HI, 21, A5XX_VFD_PERFCTR_VFD_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VFD_2_LO,
|
|
A5XX_RBBM_PERFCTR_VFD_2_HI, 22, A5XX_VFD_PERFCTR_VFD_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VFD_3_LO,
|
|
A5XX_RBBM_PERFCTR_VFD_3_HI, 23, A5XX_VFD_PERFCTR_VFD_SEL_3 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VFD_4_LO,
|
|
A5XX_RBBM_PERFCTR_VFD_4_HI, 24, A5XX_VFD_PERFCTR_VFD_SEL_4 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VFD_5_LO,
|
|
A5XX_RBBM_PERFCTR_VFD_5_HI, 25, A5XX_VFD_PERFCTR_VFD_SEL_5 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VFD_6_LO,
|
|
A5XX_RBBM_PERFCTR_VFD_6_HI, 26, A5XX_VFD_PERFCTR_VFD_SEL_6 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VFD_7_LO,
|
|
A5XX_RBBM_PERFCTR_VFD_7_HI, 27, A5XX_VFD_PERFCTR_VFD_SEL_7 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_hlsq[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_HLSQ_0_LO,
|
|
A5XX_RBBM_PERFCTR_HLSQ_0_HI, 28, A5XX_HLSQ_PERFCTR_HLSQ_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_HLSQ_1_LO,
|
|
A5XX_RBBM_PERFCTR_HLSQ_1_HI, 29, A5XX_HLSQ_PERFCTR_HLSQ_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_HLSQ_2_LO,
|
|
A5XX_RBBM_PERFCTR_HLSQ_2_HI, 30, A5XX_HLSQ_PERFCTR_HLSQ_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_HLSQ_3_LO,
|
|
A5XX_RBBM_PERFCTR_HLSQ_3_HI, 31, A5XX_HLSQ_PERFCTR_HLSQ_SEL_3 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_HLSQ_4_LO,
|
|
A5XX_RBBM_PERFCTR_HLSQ_4_HI, 32, A5XX_HLSQ_PERFCTR_HLSQ_SEL_4 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_HLSQ_5_LO,
|
|
A5XX_RBBM_PERFCTR_HLSQ_5_HI, 33, A5XX_HLSQ_PERFCTR_HLSQ_SEL_5 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_HLSQ_6_LO,
|
|
A5XX_RBBM_PERFCTR_HLSQ_6_HI, 34, A5XX_HLSQ_PERFCTR_HLSQ_SEL_6 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_HLSQ_7_LO,
|
|
A5XX_RBBM_PERFCTR_HLSQ_7_HI, 35, A5XX_HLSQ_PERFCTR_HLSQ_SEL_7 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_vpc[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VPC_0_LO,
|
|
A5XX_RBBM_PERFCTR_VPC_0_HI, 36, A5XX_VPC_PERFCTR_VPC_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VPC_1_LO,
|
|
A5XX_RBBM_PERFCTR_VPC_1_HI, 37, A5XX_VPC_PERFCTR_VPC_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VPC_2_LO,
|
|
A5XX_RBBM_PERFCTR_VPC_2_HI, 38, A5XX_VPC_PERFCTR_VPC_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VPC_3_LO,
|
|
A5XX_RBBM_PERFCTR_VPC_3_HI, 39, A5XX_VPC_PERFCTR_VPC_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_ccu[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CCU_0_LO,
|
|
A5XX_RBBM_PERFCTR_CCU_0_HI, 40, A5XX_RB_PERFCTR_CCU_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CCU_1_LO,
|
|
A5XX_RBBM_PERFCTR_CCU_1_HI, 41, A5XX_RB_PERFCTR_CCU_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CCU_2_LO,
|
|
A5XX_RBBM_PERFCTR_CCU_2_HI, 42, A5XX_RB_PERFCTR_CCU_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CCU_3_LO,
|
|
A5XX_RBBM_PERFCTR_CCU_3_HI, 43, A5XX_RB_PERFCTR_CCU_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_tse[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TSE_0_LO,
|
|
A5XX_RBBM_PERFCTR_TSE_0_HI, 44, A5XX_GRAS_PERFCTR_TSE_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TSE_1_LO,
|
|
A5XX_RBBM_PERFCTR_TSE_1_HI, 45, A5XX_GRAS_PERFCTR_TSE_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TSE_2_LO,
|
|
A5XX_RBBM_PERFCTR_TSE_2_HI, 46, A5XX_GRAS_PERFCTR_TSE_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TSE_3_LO,
|
|
A5XX_RBBM_PERFCTR_TSE_3_HI, 47, A5XX_GRAS_PERFCTR_TSE_SEL_3 },
|
|
};
|
|
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_ras[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RAS_0_LO,
|
|
A5XX_RBBM_PERFCTR_RAS_0_HI, 48, A5XX_GRAS_PERFCTR_RAS_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RAS_1_LO,
|
|
A5XX_RBBM_PERFCTR_RAS_1_HI, 49, A5XX_GRAS_PERFCTR_RAS_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RAS_2_LO,
|
|
A5XX_RBBM_PERFCTR_RAS_2_HI, 50, A5XX_GRAS_PERFCTR_RAS_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RAS_3_LO,
|
|
A5XX_RBBM_PERFCTR_RAS_3_HI, 51, A5XX_GRAS_PERFCTR_RAS_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_uche[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_UCHE_0_LO,
|
|
A5XX_RBBM_PERFCTR_UCHE_0_HI, 52, A5XX_UCHE_PERFCTR_UCHE_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_UCHE_1_LO,
|
|
A5XX_RBBM_PERFCTR_UCHE_1_HI, 53, A5XX_UCHE_PERFCTR_UCHE_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_UCHE_2_LO,
|
|
A5XX_RBBM_PERFCTR_UCHE_2_HI, 54, A5XX_UCHE_PERFCTR_UCHE_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_UCHE_3_LO,
|
|
A5XX_RBBM_PERFCTR_UCHE_3_HI, 55, A5XX_UCHE_PERFCTR_UCHE_SEL_3 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_UCHE_4_LO,
|
|
A5XX_RBBM_PERFCTR_UCHE_4_HI, 56, A5XX_UCHE_PERFCTR_UCHE_SEL_4 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_UCHE_5_LO,
|
|
A5XX_RBBM_PERFCTR_UCHE_5_HI, 57, A5XX_UCHE_PERFCTR_UCHE_SEL_5 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_UCHE_6_LO,
|
|
A5XX_RBBM_PERFCTR_UCHE_6_HI, 58, A5XX_UCHE_PERFCTR_UCHE_SEL_6 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_UCHE_7_LO,
|
|
A5XX_RBBM_PERFCTR_UCHE_7_HI, 59, A5XX_UCHE_PERFCTR_UCHE_SEL_7 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_tp[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TP_0_LO,
|
|
A5XX_RBBM_PERFCTR_TP_0_HI, 60, A5XX_TPL1_PERFCTR_TP_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TP_1_LO,
|
|
A5XX_RBBM_PERFCTR_TP_1_HI, 61, A5XX_TPL1_PERFCTR_TP_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TP_2_LO,
|
|
A5XX_RBBM_PERFCTR_TP_2_HI, 62, A5XX_TPL1_PERFCTR_TP_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TP_3_LO,
|
|
A5XX_RBBM_PERFCTR_TP_3_HI, 63, A5XX_TPL1_PERFCTR_TP_SEL_3 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TP_4_LO,
|
|
A5XX_RBBM_PERFCTR_TP_4_HI, 64, A5XX_TPL1_PERFCTR_TP_SEL_4 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TP_5_LO,
|
|
A5XX_RBBM_PERFCTR_TP_5_HI, 65, A5XX_TPL1_PERFCTR_TP_SEL_5 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TP_6_LO,
|
|
A5XX_RBBM_PERFCTR_TP_6_HI, 66, A5XX_TPL1_PERFCTR_TP_SEL_6 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_TP_7_LO,
|
|
A5XX_RBBM_PERFCTR_TP_7_HI, 67, A5XX_TPL1_PERFCTR_TP_SEL_7 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_sp[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_0_LO,
|
|
A5XX_RBBM_PERFCTR_SP_0_HI, 68, A5XX_SP_PERFCTR_SP_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_1_LO,
|
|
A5XX_RBBM_PERFCTR_SP_1_HI, 69, A5XX_SP_PERFCTR_SP_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_2_LO,
|
|
A5XX_RBBM_PERFCTR_SP_2_HI, 70, A5XX_SP_PERFCTR_SP_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_3_LO,
|
|
A5XX_RBBM_PERFCTR_SP_3_HI, 71, A5XX_SP_PERFCTR_SP_SEL_3 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_4_LO,
|
|
A5XX_RBBM_PERFCTR_SP_4_HI, 72, A5XX_SP_PERFCTR_SP_SEL_4 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_5_LO,
|
|
A5XX_RBBM_PERFCTR_SP_5_HI, 73, A5XX_SP_PERFCTR_SP_SEL_5 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_6_LO,
|
|
A5XX_RBBM_PERFCTR_SP_6_HI, 74, A5XX_SP_PERFCTR_SP_SEL_6 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_7_LO,
|
|
A5XX_RBBM_PERFCTR_SP_7_HI, 75, A5XX_SP_PERFCTR_SP_SEL_7 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_8_LO,
|
|
A5XX_RBBM_PERFCTR_SP_8_HI, 76, A5XX_SP_PERFCTR_SP_SEL_8 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_9_LO,
|
|
A5XX_RBBM_PERFCTR_SP_9_HI, 77, A5XX_SP_PERFCTR_SP_SEL_9 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_10_LO,
|
|
A5XX_RBBM_PERFCTR_SP_10_HI, 78, A5XX_SP_PERFCTR_SP_SEL_10 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_SP_11_LO,
|
|
A5XX_RBBM_PERFCTR_SP_11_HI, 79, A5XX_SP_PERFCTR_SP_SEL_11 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_rb[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RB_0_LO,
|
|
A5XX_RBBM_PERFCTR_RB_0_HI, 80, A5XX_RB_PERFCTR_RB_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RB_1_LO,
|
|
A5XX_RBBM_PERFCTR_RB_1_HI, 81, A5XX_RB_PERFCTR_RB_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RB_2_LO,
|
|
A5XX_RBBM_PERFCTR_RB_2_HI, 82, A5XX_RB_PERFCTR_RB_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RB_3_LO,
|
|
A5XX_RBBM_PERFCTR_RB_3_HI, 83, A5XX_RB_PERFCTR_RB_SEL_3 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RB_4_LO,
|
|
A5XX_RBBM_PERFCTR_RB_4_HI, 84, A5XX_RB_PERFCTR_RB_SEL_4 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RB_5_LO,
|
|
A5XX_RBBM_PERFCTR_RB_5_HI, 85, A5XX_RB_PERFCTR_RB_SEL_5 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RB_6_LO,
|
|
A5XX_RBBM_PERFCTR_RB_6_HI, 86, A5XX_RB_PERFCTR_RB_SEL_6 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RB_7_LO,
|
|
A5XX_RBBM_PERFCTR_RB_7_HI, 87, A5XX_RB_PERFCTR_RB_SEL_7 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_vsc[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VSC_0_LO,
|
|
A5XX_RBBM_PERFCTR_VSC_0_HI, 88, A5XX_VSC_PERFCTR_VSC_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_VSC_1_LO,
|
|
A5XX_RBBM_PERFCTR_VSC_1_HI, 89, A5XX_VSC_PERFCTR_VSC_SEL_1 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_lrz[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_LRZ_0_LO,
|
|
A5XX_RBBM_PERFCTR_LRZ_0_HI, 90, A5XX_GRAS_PERFCTR_LRZ_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_LRZ_1_LO,
|
|
A5XX_RBBM_PERFCTR_LRZ_1_HI, 91, A5XX_GRAS_PERFCTR_LRZ_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_LRZ_2_LO,
|
|
A5XX_RBBM_PERFCTR_LRZ_2_HI, 92, A5XX_GRAS_PERFCTR_LRZ_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_LRZ_3_LO,
|
|
A5XX_RBBM_PERFCTR_LRZ_3_HI, 93, A5XX_GRAS_PERFCTR_LRZ_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_cmp[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CMP_0_LO,
|
|
A5XX_RBBM_PERFCTR_CMP_0_HI, 94, A5XX_RB_PERFCTR_CMP_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CMP_1_LO,
|
|
A5XX_RBBM_PERFCTR_CMP_1_HI, 95, A5XX_RB_PERFCTR_CMP_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CMP_2_LO,
|
|
A5XX_RBBM_PERFCTR_CMP_2_HI, 96, A5XX_RB_PERFCTR_CMP_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_CMP_3_LO,
|
|
A5XX_RBBM_PERFCTR_CMP_3_HI, 97, A5XX_RB_PERFCTR_CMP_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_vbif[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_VBIF_PERF_CNT_LOW0,
|
|
A5XX_VBIF_PERF_CNT_HIGH0, -1, A5XX_VBIF_PERF_CNT_SEL0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_VBIF_PERF_CNT_LOW1,
|
|
A5XX_VBIF_PERF_CNT_HIGH1, -1, A5XX_VBIF_PERF_CNT_SEL1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_VBIF_PERF_CNT_LOW2,
|
|
A5XX_VBIF_PERF_CNT_HIGH2, -1, A5XX_VBIF_PERF_CNT_SEL2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_VBIF_PERF_CNT_LOW3,
|
|
A5XX_VBIF_PERF_CNT_HIGH3, -1, A5XX_VBIF_PERF_CNT_SEL3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_vbif_pwr[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_VBIF_PERF_PWR_CNT_LOW0,
|
|
A5XX_VBIF_PERF_PWR_CNT_HIGH0, -1, A5XX_VBIF_PERF_PWR_CNT_EN0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_VBIF_PERF_PWR_CNT_LOW1,
|
|
A5XX_VBIF_PERF_PWR_CNT_HIGH1, -1, A5XX_VBIF_PERF_PWR_CNT_EN1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_VBIF_PERF_PWR_CNT_LOW2,
|
|
A5XX_VBIF_PERF_PWR_CNT_HIGH2, -1, A5XX_VBIF_PERF_PWR_CNT_EN2 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_perfcounters_alwayson[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_ALWAYSON_COUNTER_LO,
|
|
A5XX_RBBM_ALWAYSON_COUNTER_HI, -1 },
|
|
};
|
|
|
|
/*
|
|
* 5XX targets don't really have physical PERFCTR_PWR registers - we emulate
|
|
* them using similar performance counters from the RBBM block. The difference
|
|
* between using this group and the RBBM group is that the RBBM counters are
|
|
* reloaded after a power collapse which is not how the PWR counters behaved on
|
|
* legacy hardware. In order to limit the disruption on the rest of the system
|
|
* we go out of our way to ensure backwards compatibility. Since RBBM counters
|
|
* are in short supply, we don't emulate PWR:0 which nobody uses - mark it as
|
|
* broken.
|
|
*/
|
|
static struct adreno_perfcount_register a5xx_perfcounters_pwr[] = {
|
|
{ KGSL_PERFCOUNTER_BROKEN, 0, 0, 0, 0, -1, 0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RBBM_PERFCTR_RBBM_0_LO,
|
|
A5XX_RBBM_PERFCTR_RBBM_0_HI, -1, 0},
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_pwrcounters_sp[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_SP_POWER_COUNTER_0_LO,
|
|
A5XX_SP_POWER_COUNTER_0_HI, -1, A5XX_SP_POWERCTR_SP_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_SP_POWER_COUNTER_1_LO,
|
|
A5XX_SP_POWER_COUNTER_1_HI, -1, A5XX_SP_POWERCTR_SP_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_SP_POWER_COUNTER_2_LO,
|
|
A5XX_SP_POWER_COUNTER_2_HI, -1, A5XX_SP_POWERCTR_SP_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_SP_POWER_COUNTER_3_LO,
|
|
A5XX_SP_POWER_COUNTER_3_HI, -1, A5XX_SP_POWERCTR_SP_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_pwrcounters_tp[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_TP_POWER_COUNTER_0_LO,
|
|
A5XX_TP_POWER_COUNTER_0_HI, -1, A5XX_TPL1_POWERCTR_TP_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_TP_POWER_COUNTER_1_LO,
|
|
A5XX_TP_POWER_COUNTER_1_HI, -1, A5XX_TPL1_POWERCTR_TP_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_TP_POWER_COUNTER_2_LO,
|
|
A5XX_TP_POWER_COUNTER_2_HI, -1, A5XX_TPL1_POWERCTR_TP_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_TP_POWER_COUNTER_3_LO,
|
|
A5XX_TP_POWER_COUNTER_3_HI, -1, A5XX_TPL1_POWERCTR_TP_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_pwrcounters_rb[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RB_POWER_COUNTER_0_LO,
|
|
A5XX_RB_POWER_COUNTER_0_HI, -1, A5XX_RB_POWERCTR_RB_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RB_POWER_COUNTER_1_LO,
|
|
A5XX_RB_POWER_COUNTER_1_HI, -1, A5XX_RB_POWERCTR_RB_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RB_POWER_COUNTER_2_LO,
|
|
A5XX_RB_POWER_COUNTER_2_HI, -1, A5XX_RB_POWERCTR_RB_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_RB_POWER_COUNTER_3_LO,
|
|
A5XX_RB_POWER_COUNTER_3_HI, -1, A5XX_RB_POWERCTR_RB_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_pwrcounters_ccu[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_CCU_POWER_COUNTER_0_LO,
|
|
A5XX_CCU_POWER_COUNTER_0_HI, -1, A5XX_RB_POWERCTR_CCU_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_CCU_POWER_COUNTER_1_LO,
|
|
A5XX_CCU_POWER_COUNTER_1_HI, -1, A5XX_RB_POWERCTR_CCU_SEL_1 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_pwrcounters_uche[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_UCHE_POWER_COUNTER_0_LO,
|
|
A5XX_UCHE_POWER_COUNTER_0_HI, -1,
|
|
A5XX_UCHE_POWERCTR_UCHE_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_UCHE_POWER_COUNTER_1_LO,
|
|
A5XX_UCHE_POWER_COUNTER_1_HI, -1,
|
|
A5XX_UCHE_POWERCTR_UCHE_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_UCHE_POWER_COUNTER_2_LO,
|
|
A5XX_UCHE_POWER_COUNTER_2_HI, -1,
|
|
A5XX_UCHE_POWERCTR_UCHE_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_UCHE_POWER_COUNTER_3_LO,
|
|
A5XX_UCHE_POWER_COUNTER_3_HI, -1,
|
|
A5XX_UCHE_POWERCTR_UCHE_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_pwrcounters_cp[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_CP_POWER_COUNTER_0_LO,
|
|
A5XX_CP_POWER_COUNTER_0_HI, -1, A5XX_CP_POWERCTR_CP_SEL_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_CP_POWER_COUNTER_1_LO,
|
|
A5XX_CP_POWER_COUNTER_1_HI, -1, A5XX_CP_POWERCTR_CP_SEL_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_CP_POWER_COUNTER_2_LO,
|
|
A5XX_CP_POWER_COUNTER_2_HI, -1, A5XX_CP_POWERCTR_CP_SEL_2 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_CP_POWER_COUNTER_3_LO,
|
|
A5XX_CP_POWER_COUNTER_3_HI, -1, A5XX_CP_POWERCTR_CP_SEL_3 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_pwrcounters_gpmu[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_GPMU_POWER_COUNTER_0_LO,
|
|
A5XX_GPMU_POWER_COUNTER_0_HI, -1,
|
|
A5XX_GPMU_POWER_COUNTER_SELECT_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_GPMU_POWER_COUNTER_1_LO,
|
|
A5XX_GPMU_POWER_COUNTER_1_HI, -1,
|
|
A5XX_GPMU_POWER_COUNTER_SELECT_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_GPMU_POWER_COUNTER_2_LO,
|
|
A5XX_GPMU_POWER_COUNTER_2_HI, -1,
|
|
A5XX_GPMU_POWER_COUNTER_SELECT_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_GPMU_POWER_COUNTER_3_LO,
|
|
A5XX_GPMU_POWER_COUNTER_3_HI, -1,
|
|
A5XX_GPMU_POWER_COUNTER_SELECT_0 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_GPMU_POWER_COUNTER_4_LO,
|
|
A5XX_GPMU_POWER_COUNTER_4_HI, -1,
|
|
A5XX_GPMU_POWER_COUNTER_SELECT_1 },
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_GPMU_POWER_COUNTER_5_LO,
|
|
A5XX_GPMU_POWER_COUNTER_5_HI, -1,
|
|
A5XX_GPMU_POWER_COUNTER_SELECT_1 },
|
|
};
|
|
|
|
static struct adreno_perfcount_register a5xx_pwrcounters_alwayson[] = {
|
|
{ KGSL_PERFCOUNTER_NOT_USED, 0, 0, A5XX_GPMU_ALWAYS_ON_COUNTER_LO,
|
|
A5XX_GPMU_ALWAYS_ON_COUNTER_HI, -1 },
|
|
};
|
|
|
|
#define A5XX_PERFCOUNTER_GROUP(offset, name) \
|
|
ADRENO_PERFCOUNTER_GROUP(a5xx, offset, name)
|
|
|
|
#define A5XX_PERFCOUNTER_GROUP_FLAGS(offset, name, flags) \
|
|
ADRENO_PERFCOUNTER_GROUP_FLAGS(a5xx, offset, name, flags)
|
|
|
|
#define A5XX_POWER_COUNTER_GROUP(offset, name) \
|
|
ADRENO_POWER_COUNTER_GROUP(a5xx, offset, name)
|
|
|
|
static struct adreno_perfcount_group a5xx_perfcounter_groups
|
|
[KGSL_PERFCOUNTER_GROUP_MAX] = {
|
|
A5XX_PERFCOUNTER_GROUP(CP, cp),
|
|
A5XX_PERFCOUNTER_GROUP(RBBM, rbbm),
|
|
A5XX_PERFCOUNTER_GROUP(PC, pc),
|
|
A5XX_PERFCOUNTER_GROUP(VFD, vfd),
|
|
A5XX_PERFCOUNTER_GROUP(HLSQ, hlsq),
|
|
A5XX_PERFCOUNTER_GROUP(VPC, vpc),
|
|
A5XX_PERFCOUNTER_GROUP(CCU, ccu),
|
|
A5XX_PERFCOUNTER_GROUP(CMP, cmp),
|
|
A5XX_PERFCOUNTER_GROUP(TSE, tse),
|
|
A5XX_PERFCOUNTER_GROUP(RAS, ras),
|
|
A5XX_PERFCOUNTER_GROUP(LRZ, lrz),
|
|
A5XX_PERFCOUNTER_GROUP(UCHE, uche),
|
|
A5XX_PERFCOUNTER_GROUP(TP, tp),
|
|
A5XX_PERFCOUNTER_GROUP(SP, sp),
|
|
A5XX_PERFCOUNTER_GROUP(RB, rb),
|
|
A5XX_PERFCOUNTER_GROUP(VSC, vsc),
|
|
A5XX_PERFCOUNTER_GROUP_FLAGS(PWR, pwr,
|
|
ADRENO_PERFCOUNTER_GROUP_FIXED),
|
|
A5XX_PERFCOUNTER_GROUP(VBIF, vbif),
|
|
A5XX_PERFCOUNTER_GROUP_FLAGS(VBIF_PWR, vbif_pwr,
|
|
ADRENO_PERFCOUNTER_GROUP_FIXED),
|
|
A5XX_PERFCOUNTER_GROUP_FLAGS(ALWAYSON, alwayson,
|
|
ADRENO_PERFCOUNTER_GROUP_FIXED),
|
|
A5XX_POWER_COUNTER_GROUP(SP, sp),
|
|
A5XX_POWER_COUNTER_GROUP(TP, tp),
|
|
A5XX_POWER_COUNTER_GROUP(RB, rb),
|
|
A5XX_POWER_COUNTER_GROUP(CCU, ccu),
|
|
A5XX_POWER_COUNTER_GROUP(UCHE, uche),
|
|
A5XX_POWER_COUNTER_GROUP(CP, cp),
|
|
A5XX_POWER_COUNTER_GROUP(GPMU, gpmu),
|
|
A5XX_POWER_COUNTER_GROUP(ALWAYSON, alwayson),
|
|
};
|
|
|
|
static struct adreno_perfcounters a5xx_perfcounters = {
|
|
a5xx_perfcounter_groups,
|
|
ARRAY_SIZE(a5xx_perfcounter_groups),
|
|
};
|
|
|
|
static struct adreno_ft_perf_counters a5xx_ft_perf_counters[] = {
|
|
{KGSL_PERFCOUNTER_GROUP_SP, A5XX_SP_ALU_ACTIVE_CYCLES},
|
|
{KGSL_PERFCOUNTER_GROUP_SP, A5XX_SP0_ICL1_MISSES},
|
|
{KGSL_PERFCOUNTER_GROUP_SP, A5XX_SP_FS_CFLOW_INSTRUCTIONS},
|
|
{KGSL_PERFCOUNTER_GROUP_TSE, A5XX_TSE_INPUT_PRIM_NUM},
|
|
};
|
|
|
|
static unsigned int a5xx_int_bits[ADRENO_INT_BITS_MAX] = {
|
|
ADRENO_INT_DEFINE(ADRENO_INT_RBBM_AHB_ERROR, A5XX_INT_RBBM_AHB_ERROR),
|
|
};
|
|
|
|
/* Register offset defines for A5XX, in order of enum adreno_regs */
|
|
static unsigned int a5xx_register_offsets[ADRENO_REG_REGISTER_MAX] = {
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_WFI_PEND_CTR, A5XX_CP_WFI_PEND_CTR),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_RB_BASE, A5XX_CP_RB_BASE),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_RB_BASE_HI, A5XX_CP_RB_BASE_HI),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_RB_RPTR_ADDR_LO,
|
|
A5XX_CP_RB_RPTR_ADDR_LO),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_RB_RPTR_ADDR_HI,
|
|
A5XX_CP_RB_RPTR_ADDR_HI),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_RB_RPTR, A5XX_CP_RB_RPTR),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_RB_WPTR, A5XX_CP_RB_WPTR),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_CNTL, A5XX_CP_CNTL),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_ME_CNTL, A5XX_CP_ME_CNTL),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_RB_CNTL, A5XX_CP_RB_CNTL),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_IB1_BASE, A5XX_CP_IB1_BASE),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_IB1_BASE_HI, A5XX_CP_IB1_BASE_HI),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_IB1_BUFSZ, A5XX_CP_IB1_BUFSZ),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_IB2_BASE, A5XX_CP_IB2_BASE),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_IB2_BASE_HI, A5XX_CP_IB2_BASE_HI),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_IB2_BUFSZ, A5XX_CP_IB2_BUFSZ),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_ROQ_ADDR, A5XX_CP_ROQ_DBG_ADDR),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_ROQ_DATA, A5XX_CP_ROQ_DBG_DATA),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_MERCIU_ADDR, A5XX_CP_MERCIU_DBG_ADDR),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_MERCIU_DATA, A5XX_CP_MERCIU_DBG_DATA_1),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_MERCIU_DATA2,
|
|
A5XX_CP_MERCIU_DBG_DATA_2),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_MEQ_ADDR, A5XX_CP_MEQ_DBG_ADDR),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_MEQ_DATA, A5XX_CP_MEQ_DBG_DATA),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_PROTECT_REG_0, A5XX_CP_PROTECT_REG_0),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_HW_FAULT, A5XX_CP_HW_FAULT),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_PREEMPT, A5XX_CP_CONTEXT_SWITCH_CNTL),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_PREEMPT_DEBUG, ADRENO_REG_SKIP),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_PREEMPT_DISABLE, ADRENO_REG_SKIP),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_CONTEXT_SWITCH_SMMU_INFO_LO,
|
|
A5XX_CP_CONTEXT_SWITCH_SMMU_INFO_LO),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_CP_CONTEXT_SWITCH_SMMU_INFO_HI,
|
|
A5XX_CP_CONTEXT_SWITCH_SMMU_INFO_HI),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_STATUS, A5XX_RBBM_STATUS),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_STATUS3, A5XX_RBBM_STATUS3),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_PERFCTR_CTL, A5XX_RBBM_PERFCTR_CNTL),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_PERFCTR_LOAD_CMD0,
|
|
A5XX_RBBM_PERFCTR_LOAD_CMD0),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_PERFCTR_LOAD_CMD1,
|
|
A5XX_RBBM_PERFCTR_LOAD_CMD1),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_PERFCTR_LOAD_CMD2,
|
|
A5XX_RBBM_PERFCTR_LOAD_CMD2),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_PERFCTR_LOAD_CMD3,
|
|
A5XX_RBBM_PERFCTR_LOAD_CMD3),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_INT_0_MASK, A5XX_RBBM_INT_0_MASK),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_INT_0_STATUS, A5XX_RBBM_INT_0_STATUS),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_CLOCK_CTL, A5XX_RBBM_CLOCK_CNTL),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_INT_CLEAR_CMD,
|
|
A5XX_RBBM_INT_CLEAR_CMD),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_SW_RESET_CMD, A5XX_RBBM_SW_RESET_CMD),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_BLOCK_SW_RESET_CMD,
|
|
A5XX_RBBM_BLOCK_SW_RESET_CMD),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_BLOCK_SW_RESET_CMD2,
|
|
A5XX_RBBM_BLOCK_SW_RESET_CMD2),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_UCHE_INVALIDATE0, A5XX_UCHE_INVALIDATE0),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_PERFCTR_RBBM_0_LO,
|
|
A5XX_RBBM_PERFCTR_RBBM_0_LO),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_PERFCTR_RBBM_0_HI,
|
|
A5XX_RBBM_PERFCTR_RBBM_0_HI),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_PERFCTR_LOAD_VALUE_LO,
|
|
A5XX_RBBM_PERFCTR_LOAD_VALUE_LO),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_PERFCTR_LOAD_VALUE_HI,
|
|
A5XX_RBBM_PERFCTR_LOAD_VALUE_HI),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_SECVID_TRUST_CONTROL,
|
|
A5XX_RBBM_SECVID_TRUST_CNTL),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_SECVID_TRUST_CONFIG,
|
|
A5XX_RBBM_SECVID_TRUST_CONFIG),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_SECVID_TSB_CONTROL,
|
|
A5XX_RBBM_SECVID_TSB_CNTL),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_SECVID_TSB_TRUSTED_BASE,
|
|
A5XX_RBBM_SECVID_TSB_TRUSTED_BASE_LO),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_SECVID_TSB_TRUSTED_BASE_HI,
|
|
A5XX_RBBM_SECVID_TSB_TRUSTED_BASE_HI),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_SECVID_TSB_TRUSTED_SIZE,
|
|
A5XX_RBBM_SECVID_TSB_TRUSTED_SIZE),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_ALWAYSON_COUNTER_LO,
|
|
A5XX_RBBM_ALWAYSON_COUNTER_LO),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_RBBM_ALWAYSON_COUNTER_HI,
|
|
A5XX_RBBM_ALWAYSON_COUNTER_HI),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_VBIF_XIN_HALT_CTRL0,
|
|
A5XX_VBIF_XIN_HALT_CTRL0),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_VBIF_XIN_HALT_CTRL1,
|
|
A5XX_VBIF_XIN_HALT_CTRL1),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_VBIF_VERSION,
|
|
A5XX_VBIF_VERSION),
|
|
ADRENO_REG_DEFINE(ADRENO_REG_GPMU_POWER_COUNTER_ENABLE,
|
|
A5XX_GPMU_POWER_COUNTER_ENABLE),
|
|
};
|
|
|
|
static const struct adreno_reg_offsets a5xx_reg_offsets = {
|
|
.offsets = a5xx_register_offsets,
|
|
.offset_0 = ADRENO_REG_REGISTER_MAX,
|
|
};
|
|
|
|
static void a5xx_cp_hw_err_callback(struct adreno_device *adreno_dev, int bit)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
unsigned int status1, status2;
|
|
|
|
kgsl_regread(device, A5XX_CP_INTERRUPT_STATUS, &status1);
|
|
|
|
if (status1 & BIT(A5XX_CP_OPCODE_ERROR)) {
|
|
unsigned int val;
|
|
|
|
kgsl_regwrite(device, A5XX_CP_PFP_STAT_ADDR, 0);
|
|
|
|
/*
|
|
* A5XX_CP_PFP_STAT_DATA is indexed, so read it twice to get the
|
|
* value we want
|
|
*/
|
|
kgsl_regread(device, A5XX_CP_PFP_STAT_DATA, &val);
|
|
kgsl_regread(device, A5XX_CP_PFP_STAT_DATA, &val);
|
|
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"ringbuffer opcode error | possible opcode=0x%8.8X\n",
|
|
val);
|
|
}
|
|
if (status1 & BIT(A5XX_CP_RESERVED_BIT_ERROR))
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"ringbuffer reserved bit error interrupt\n");
|
|
if (status1 & BIT(A5XX_CP_HW_FAULT_ERROR)) {
|
|
kgsl_regread(device, A5XX_CP_HW_FAULT, &status2);
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"CP | Ringbuffer HW fault | status=%x\n",
|
|
status2);
|
|
}
|
|
if (status1 & BIT(A5XX_CP_DMA_ERROR))
|
|
KGSL_DRV_CRIT_RATELIMIT(device, "CP | DMA error\n");
|
|
if (status1 & BIT(A5XX_CP_REGISTER_PROTECTION_ERROR)) {
|
|
kgsl_regread(device, A5XX_CP_PROTECT_STATUS, &status2);
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"CP | Protected mode error| %s | addr=%x | status=%x\n",
|
|
status2 & (1 << 24) ? "WRITE" : "READ",
|
|
(status2 & 0xFFFFF) >> 2, status2);
|
|
}
|
|
if (status1 & BIT(A5XX_CP_AHB_ERROR)) {
|
|
kgsl_regread(device, A5XX_CP_AHB_FAULT, &status2);
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"ringbuffer AHB error interrupt | status=%x\n",
|
|
status2);
|
|
}
|
|
}
|
|
|
|
static void a5xx_err_callback(struct adreno_device *adreno_dev, int bit)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
unsigned int reg;
|
|
|
|
switch (bit) {
|
|
case A5XX_INT_RBBM_AHB_ERROR: {
|
|
kgsl_regread(device, A5XX_RBBM_AHB_ERROR_STATUS, ®);
|
|
|
|
/*
|
|
* Return the word address of the erroring register so that it
|
|
* matches the register specification
|
|
*/
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"RBBM | AHB bus error | %s | addr=%x | ports=%x:%x\n",
|
|
reg & (1 << 28) ? "WRITE" : "READ",
|
|
(reg & 0xFFFFF) >> 2, (reg >> 20) & 0x3,
|
|
(reg >> 24) & 0xF);
|
|
|
|
/* Clear the error */
|
|
kgsl_regwrite(device, A5XX_RBBM_AHB_CMD, (1 << 4));
|
|
break;
|
|
}
|
|
case A5XX_INT_RBBM_TRANSFER_TIMEOUT:
|
|
KGSL_DRV_CRIT_RATELIMIT(device, "RBBM: AHB transfer timeout\n");
|
|
break;
|
|
case A5XX_INT_RBBM_ME_MS_TIMEOUT:
|
|
kgsl_regread(device, A5XX_RBBM_AHB_ME_SPLIT_STATUS, ®);
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"RBBM | ME master split timeout | status=%x\n", reg);
|
|
break;
|
|
case A5XX_INT_RBBM_PFP_MS_TIMEOUT:
|
|
kgsl_regread(device, A5XX_RBBM_AHB_PFP_SPLIT_STATUS, ®);
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"RBBM | PFP master split timeout | status=%x\n", reg);
|
|
break;
|
|
case A5XX_INT_RBBM_ETS_MS_TIMEOUT:
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"RBBM: ME master split timeout\n");
|
|
break;
|
|
case A5XX_INT_RBBM_ATB_ASYNC_OVERFLOW:
|
|
KGSL_DRV_CRIT_RATELIMIT(device, "RBBM: ATB ASYNC overflow\n");
|
|
break;
|
|
case A5XX_INT_RBBM_ATB_BUS_OVERFLOW:
|
|
KGSL_DRV_CRIT_RATELIMIT(device, "RBBM: ATB bus overflow\n");
|
|
break;
|
|
case A5XX_INT_UCHE_OOB_ACCESS:
|
|
KGSL_DRV_CRIT_RATELIMIT(device, "UCHE: Out of bounds access\n");
|
|
break;
|
|
case A5XX_INT_UCHE_TRAP_INTR:
|
|
KGSL_DRV_CRIT_RATELIMIT(device, "UCHE: Trap interrupt\n");
|
|
break;
|
|
case A5XX_INT_GPMU_VOLTAGE_DROOP:
|
|
KGSL_DRV_CRIT_RATELIMIT(device, "GPMU: Voltage droop\n");
|
|
break;
|
|
default:
|
|
KGSL_DRV_CRIT_RATELIMIT(device, "Unknown interrupt %d\n", bit);
|
|
}
|
|
}
|
|
|
|
static void a5xx_irq_storm_worker(struct work_struct *work)
|
|
{
|
|
struct adreno_device *adreno_dev = container_of(work,
|
|
struct adreno_device, irq_storm_work);
|
|
struct kgsl_device *device = &adreno_dev->dev;
|
|
struct adreno_gpudev *gpudev = ADRENO_GPU_DEVICE(adreno_dev);
|
|
unsigned int status;
|
|
|
|
mutex_lock(&device->mutex);
|
|
|
|
/* Wait for the storm to clear up */
|
|
do {
|
|
adreno_writereg(adreno_dev, ADRENO_REG_RBBM_INT_CLEAR_CMD,
|
|
BIT(A5XX_INT_CP_CACHE_FLUSH_TS));
|
|
adreno_readreg(adreno_dev, ADRENO_REG_RBBM_INT_0_STATUS,
|
|
&status);
|
|
} while (status & BIT(A5XX_INT_CP_CACHE_FLUSH_TS));
|
|
|
|
/* Re-enable the interrupt bit in the mask */
|
|
gpudev->irq->mask |= BIT(A5XX_INT_CP_CACHE_FLUSH_TS);
|
|
adreno_writereg(adreno_dev, ADRENO_REG_RBBM_INT_0_MASK,
|
|
gpudev->irq->mask);
|
|
clear_bit(ADRENO_DEVICE_CACHE_FLUSH_TS_SUSPENDED, &adreno_dev->priv);
|
|
|
|
KGSL_DRV_WARN(device, "Re-enabled A5XX_INT_CP_CACHE_FLUSH_TS");
|
|
mutex_unlock(&device->mutex);
|
|
|
|
/* Reschedule just to make sure everything retires */
|
|
adreno_dispatcher_schedule(device);
|
|
}
|
|
|
|
static void a5xx_cp_callback(struct adreno_device *adreno_dev, int bit)
|
|
{
|
|
struct kgsl_device *device = &adreno_dev->dev;
|
|
unsigned int cur;
|
|
static unsigned int count;
|
|
static unsigned int prev;
|
|
|
|
if (test_bit(ADRENO_DEVICE_CACHE_FLUSH_TS_SUSPENDED, &adreno_dev->priv))
|
|
return;
|
|
|
|
kgsl_sharedmem_readl(&device->memstore, &cur,
|
|
KGSL_MEMSTORE_OFFSET(KGSL_MEMSTORE_GLOBAL,
|
|
ref_wait_ts));
|
|
|
|
/*
|
|
* prev holds a previously read value
|
|
* from memory. It should be changed by the GPU with every
|
|
* interrupt. If the value we know about and the value we just
|
|
* read are the same, then we are likely in a storm.
|
|
* If this happens twice, disable the interrupt in the mask
|
|
* so the dispatcher can take care of the issue. It is then
|
|
* up to the dispatcher to re-enable the mask once all work
|
|
* is done and the storm has ended.
|
|
*/
|
|
if (prev == cur) {
|
|
count++;
|
|
if (count == 2) {
|
|
struct adreno_gpudev *gpudev =
|
|
ADRENO_GPU_DEVICE(adreno_dev);
|
|
|
|
/* disable interrupt from the mask */
|
|
set_bit(ADRENO_DEVICE_CACHE_FLUSH_TS_SUSPENDED,
|
|
&adreno_dev->priv);
|
|
gpudev->irq->mask &= ~BIT(A5XX_INT_CP_CACHE_FLUSH_TS);
|
|
adreno_writereg(adreno_dev, ADRENO_REG_RBBM_INT_0_MASK,
|
|
gpudev->irq->mask);
|
|
|
|
kgsl_schedule_work(&adreno_dev->irq_storm_work);
|
|
|
|
return;
|
|
}
|
|
} else {
|
|
count = 0;
|
|
prev = cur;
|
|
}
|
|
|
|
a5xx_preemption_trigger(adreno_dev);
|
|
adreno_dispatcher_schedule(device);
|
|
}
|
|
|
|
static const char *gpmu_int_msg[32] = {
|
|
[FW_INTR_INFO] = "FW_INTR_INFO",
|
|
[LLM_ACK_ERR_INTR] = "LLM_ACK_ERR_INTR",
|
|
[ISENS_TRIM_ERR_INTR] = "ISENS_TRIM_ERR_INTR",
|
|
[ISENS_ERR_INTR] = "ISENS_ERR_INTR",
|
|
[ISENS_IDLE_ERR_INTR] = "ISENS_IDLE_ERR_INTR",
|
|
[ISENS_PWR_ON_ERR_INTR] = "ISENS_PWR_ON_ERR_INTR",
|
|
[6 ... 30] = "",
|
|
[WDOG_EXPITED] = "WDOG_EXPITED"};
|
|
|
|
static void a5xx_gpmu_int_callback(struct adreno_device *adreno_dev, int bit)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
unsigned int reg, i;
|
|
|
|
kgsl_regread(device, A5XX_GPMU_RBBM_INTR_INFO, ®);
|
|
|
|
if (reg & (~VALID_GPMU_IRQ)) {
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"GPMU: Unknown IRQ mask 0x%08lx in 0x%08x\n",
|
|
reg & (~VALID_GPMU_IRQ), reg);
|
|
}
|
|
|
|
for (i = 0; i < 32; i++)
|
|
switch (reg & BIT(i)) {
|
|
case BIT(WDOG_EXPITED):
|
|
if (test_and_clear_bit(ADRENO_DEVICE_GPMU_INITIALIZED,
|
|
&adreno_dev->priv)) {
|
|
/* Stop GPMU */
|
|
kgsl_regwrite(device,
|
|
A5XX_GPMU_CM3_SYSRESET, 1);
|
|
kgsl_schedule_work(&adreno_dev->gpmu_work);
|
|
}
|
|
/* fallthrough */
|
|
case BIT(FW_INTR_INFO):
|
|
case BIT(LLM_ACK_ERR_INTR):
|
|
case BIT(ISENS_TRIM_ERR_INTR):
|
|
case BIT(ISENS_ERR_INTR):
|
|
case BIT(ISENS_IDLE_ERR_INTR):
|
|
case BIT(ISENS_PWR_ON_ERR_INTR):
|
|
KGSL_DRV_CRIT_RATELIMIT(device,
|
|
"GPMU: interrupt %s(%08lx)\n",
|
|
gpmu_int_msg[i],
|
|
BIT(i));
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* a5xx_gpc_err_int_callback() - Isr for GPC error interrupts
|
|
* @adreno_dev: Pointer to device
|
|
* @bit: Interrupt bit
|
|
*/
|
|
static void a5xx_gpc_err_int_callback(struct adreno_device *adreno_dev, int bit)
|
|
{
|
|
struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
|
|
|
|
/*
|
|
* GPC error is typically the result of mistake SW programming.
|
|
* Force GPU fault for this interrupt so that we can debug it
|
|
* with help of register dump.
|
|
*/
|
|
|
|
KGSL_DRV_CRIT_RATELIMIT(device, "RBBM: GPC error\n");
|
|
adreno_irqctrl(adreno_dev, 0);
|
|
|
|
/* Trigger a fault in the dispatcher - this will effect a restart */
|
|
adreno_set_gpu_fault(adreno_dev, ADRENO_SOFT_FAULT);
|
|
adreno_dispatcher_schedule(device);
|
|
}
|
|
|
|
#define A5XX_INT_MASK \
|
|
((1 << A5XX_INT_RBBM_AHB_ERROR) | \
|
|
(1 << A5XX_INT_RBBM_TRANSFER_TIMEOUT) | \
|
|
(1 << A5XX_INT_RBBM_ME_MS_TIMEOUT) | \
|
|
(1 << A5XX_INT_RBBM_PFP_MS_TIMEOUT) | \
|
|
(1 << A5XX_INT_RBBM_ETS_MS_TIMEOUT) | \
|
|
(1 << A5XX_INT_RBBM_ATB_ASYNC_OVERFLOW) | \
|
|
(1 << A5XX_INT_RBBM_GPC_ERROR) | \
|
|
(1 << A5XX_INT_CP_HW_ERROR) | \
|
|
(1 << A5XX_INT_CP_CACHE_FLUSH_TS) | \
|
|
(1 << A5XX_INT_RBBM_ATB_BUS_OVERFLOW) | \
|
|
(1 << A5XX_INT_UCHE_OOB_ACCESS) | \
|
|
(1 << A5XX_INT_UCHE_TRAP_INTR) | \
|
|
(1 << A5XX_INT_CP_SW) | \
|
|
(1 << A5XX_INT_GPMU_FIRMWARE) | \
|
|
(1 << A5XX_INT_GPMU_VOLTAGE_DROOP))
|
|
|
|
|
|
static struct adreno_irq_funcs a5xx_irq_funcs[32] = {
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 0 - RBBM_GPU_IDLE */
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback), /* 1 - RBBM_AHB_ERROR */
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback), /* 2 - RBBM_TRANSFER_TIMEOUT */
|
|
/* 3 - RBBM_ME_MASTER_SPLIT_TIMEOUT */
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback),
|
|
/* 4 - RBBM_PFP_MASTER_SPLIT_TIMEOUT */
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback),
|
|
/* 5 - RBBM_ETS_MASTER_SPLIT_TIMEOUT */
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback),
|
|
/* 6 - RBBM_ATB_ASYNC_OVERFLOW */
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback),
|
|
ADRENO_IRQ_CALLBACK(a5xx_gpc_err_int_callback), /* 7 - GPC_ERR */
|
|
ADRENO_IRQ_CALLBACK(a5xx_preempt_callback),/* 8 - CP_SW */
|
|
ADRENO_IRQ_CALLBACK(a5xx_cp_hw_err_callback), /* 9 - CP_HW_ERROR */
|
|
/* 10 - CP_CCU_FLUSH_DEPTH_TS */
|
|
ADRENO_IRQ_CALLBACK(NULL),
|
|
/* 11 - CP_CCU_FLUSH_COLOR_TS */
|
|
ADRENO_IRQ_CALLBACK(NULL),
|
|
/* 12 - CP_CCU_RESOLVE_TS */
|
|
ADRENO_IRQ_CALLBACK(NULL),
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 13 - CP_IB2_INT */
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 14 - CP_IB1_INT */
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 15 - CP_RB_INT */
|
|
/* 16 - CCP_UNUSED_1 */
|
|
ADRENO_IRQ_CALLBACK(NULL),
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 17 - CP_RB_DONE_TS */
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 18 - CP_WT_DONE_TS */
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 19 - UNKNOWN_1 */
|
|
ADRENO_IRQ_CALLBACK(a5xx_cp_callback), /* 20 - CP_CACHE_FLUSH_TS */
|
|
/* 21 - UNUSED_2 */
|
|
ADRENO_IRQ_CALLBACK(NULL),
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback), /* 22 - RBBM_ATB_BUS_OVERFLOW */
|
|
/* 23 - MISC_HANG_DETECT */
|
|
ADRENO_IRQ_CALLBACK(adreno_hang_int_callback),
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback), /* 24 - UCHE_OOB_ACCESS */
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback), /* 25 - UCHE_TRAP_INTR */
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 26 - DEBBUS_INTR_0 */
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 27 - DEBBUS_INTR_1 */
|
|
ADRENO_IRQ_CALLBACK(a5xx_err_callback), /* 28 - GPMU_VOLTAGE_DROOP */
|
|
ADRENO_IRQ_CALLBACK(a5xx_gpmu_int_callback), /* 29 - GPMU_FIRMWARE */
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 30 - ISDB_CPU_IRQ */
|
|
ADRENO_IRQ_CALLBACK(NULL), /* 31 - ISDB_UNDER_DEBUG */
|
|
};
|
|
|
|
static struct adreno_irq a5xx_irq = {
|
|
.funcs = a5xx_irq_funcs,
|
|
.mask = A5XX_INT_MASK,
|
|
};
|
|
|
|
/*
|
|
* Default size for CP queues for A5xx targets. You must
|
|
* overwrite these value in platform_setup function for
|
|
* A5xx derivatives if size differs.
|
|
*/
|
|
static struct adreno_snapshot_sizes a5xx_snap_sizes = {
|
|
.cp_pfp = 36,
|
|
.cp_me = 29,
|
|
.cp_meq = 64,
|
|
.cp_merciu = 64,
|
|
.roq = 512,
|
|
};
|
|
|
|
static struct adreno_snapshot_data a5xx_snapshot_data = {
|
|
.sect_sizes = &a5xx_snap_sizes,
|
|
};
|
|
|
|
static struct adreno_coresight_register a5xx_coresight_registers[] = {
|
|
{ A5XX_RBBM_CFG_DBGBUS_SEL_A },
|
|
{ A5XX_RBBM_CFG_DBGBUS_SEL_B },
|
|
{ A5XX_RBBM_CFG_DBGBUS_SEL_C },
|
|
{ A5XX_RBBM_CFG_DBGBUS_SEL_D },
|
|
{ A5XX_RBBM_CFG_DBGBUS_CNTLT },
|
|
{ A5XX_RBBM_CFG_DBGBUS_CNTLM },
|
|
{ A5XX_RBBM_CFG_DBGBUS_OPL },
|
|
{ A5XX_RBBM_CFG_DBGBUS_OPE },
|
|
{ A5XX_RBBM_CFG_DBGBUS_IVTL_0 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_IVTL_1 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_IVTL_2 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_IVTL_3 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MASKL_0 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MASKL_1 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MASKL_2 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MASKL_3 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_BYTEL_0 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_BYTEL_1 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_IVTE_0 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_IVTE_1 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_IVTE_2 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_IVTE_3 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MASKE_0 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MASKE_1 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MASKE_2 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MASKE_3 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_NIBBLEE },
|
|
{ A5XX_RBBM_CFG_DBGBUS_PTRC0 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_PTRC1 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_LOADREG },
|
|
{ A5XX_RBBM_CFG_DBGBUS_IDX },
|
|
{ A5XX_RBBM_CFG_DBGBUS_CLRC },
|
|
{ A5XX_RBBM_CFG_DBGBUS_LOADIVT },
|
|
{ A5XX_RBBM_CFG_DBGBUS_EVENT_LOGIC },
|
|
{ A5XX_RBBM_CFG_DBGBUS_OVER },
|
|
{ A5XX_RBBM_CFG_DBGBUS_COUNT0 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_COUNT1 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_COUNT2 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_COUNT3 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_COUNT4 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_COUNT5 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_TRACE_ADDR },
|
|
{ A5XX_RBBM_CFG_DBGBUS_TRACE_BUF0 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_TRACE_BUF1 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_TRACE_BUF2 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_TRACE_BUF3 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_TRACE_BUF4 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MISR0 },
|
|
{ A5XX_RBBM_CFG_DBGBUS_MISR1 },
|
|
{ A5XX_RBBM_AHB_DBG_CNTL },
|
|
{ A5XX_RBBM_READ_AHB_THROUGH_DBG },
|
|
{ A5XX_RBBM_DBG_LO_HI_GPIO },
|
|
{ A5XX_RBBM_EXT_TRACE_BUS_CNTL },
|
|
{ A5XX_RBBM_EXT_VBIF_DBG_CNTL },
|
|
};
|
|
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_sel_a, &a5xx_coresight_registers[0]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_sel_b, &a5xx_coresight_registers[1]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_sel_c, &a5xx_coresight_registers[2]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_sel_d, &a5xx_coresight_registers[3]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_cntlt, &a5xx_coresight_registers[4]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_cntlm, &a5xx_coresight_registers[5]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_opl, &a5xx_coresight_registers[6]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ope, &a5xx_coresight_registers[7]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ivtl_0, &a5xx_coresight_registers[8]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ivtl_1, &a5xx_coresight_registers[9]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ivtl_2, &a5xx_coresight_registers[10]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ivtl_3, &a5xx_coresight_registers[11]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_maskl_0, &a5xx_coresight_registers[12]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_maskl_1, &a5xx_coresight_registers[13]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_maskl_2, &a5xx_coresight_registers[14]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_maskl_3, &a5xx_coresight_registers[15]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_bytel_0, &a5xx_coresight_registers[16]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_bytel_1, &a5xx_coresight_registers[17]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ivte_0, &a5xx_coresight_registers[18]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ivte_1, &a5xx_coresight_registers[19]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ivte_2, &a5xx_coresight_registers[20]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ivte_3, &a5xx_coresight_registers[21]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_maske_0, &a5xx_coresight_registers[22]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_maske_1, &a5xx_coresight_registers[23]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_maske_2, &a5xx_coresight_registers[24]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_maske_3, &a5xx_coresight_registers[25]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_nibblee, &a5xx_coresight_registers[26]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ptrc0, &a5xx_coresight_registers[27]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_ptrc1, &a5xx_coresight_registers[28]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_loadreg, &a5xx_coresight_registers[29]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_idx, &a5xx_coresight_registers[30]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_clrc, &a5xx_coresight_registers[31]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_loadivt, &a5xx_coresight_registers[32]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_event_logic,
|
|
&a5xx_coresight_registers[33]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_over, &a5xx_coresight_registers[34]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_count0, &a5xx_coresight_registers[35]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_count1, &a5xx_coresight_registers[36]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_count2, &a5xx_coresight_registers[37]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_count3, &a5xx_coresight_registers[38]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_count4, &a5xx_coresight_registers[39]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_count5, &a5xx_coresight_registers[40]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_trace_addr,
|
|
&a5xx_coresight_registers[41]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_trace_buf0,
|
|
&a5xx_coresight_registers[42]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_trace_buf1,
|
|
&a5xx_coresight_registers[43]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_trace_buf2,
|
|
&a5xx_coresight_registers[44]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_trace_buf3,
|
|
&a5xx_coresight_registers[45]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_trace_buf4,
|
|
&a5xx_coresight_registers[46]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_misr0, &a5xx_coresight_registers[47]);
|
|
static ADRENO_CORESIGHT_ATTR(cfg_dbgbus_misr1, &a5xx_coresight_registers[48]);
|
|
static ADRENO_CORESIGHT_ATTR(ahb_dbg_cntl, &a5xx_coresight_registers[49]);
|
|
static ADRENO_CORESIGHT_ATTR(read_ahb_through_dbg,
|
|
&a5xx_coresight_registers[50]);
|
|
static ADRENO_CORESIGHT_ATTR(dbg_lo_hi_gpio, &a5xx_coresight_registers[51]);
|
|
static ADRENO_CORESIGHT_ATTR(ext_trace_bus_cntl, &a5xx_coresight_registers[52]);
|
|
static ADRENO_CORESIGHT_ATTR(ext_vbif_dbg_cntl, &a5xx_coresight_registers[53]);
|
|
|
|
static struct attribute *a5xx_coresight_attrs[] = {
|
|
&coresight_attr_cfg_dbgbus_sel_a.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_sel_b.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_sel_c.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_sel_d.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_cntlt.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_cntlm.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_opl.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ope.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ivtl_0.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ivtl_1.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ivtl_2.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ivtl_3.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_maskl_0.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_maskl_1.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_maskl_2.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_maskl_3.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_bytel_0.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_bytel_1.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ivte_0.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ivte_1.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ivte_2.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ivte_3.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_maske_0.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_maske_1.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_maske_2.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_maske_3.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_nibblee.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ptrc0.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_ptrc1.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_loadreg.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_idx.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_clrc.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_loadivt.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_event_logic.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_over.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_count0.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_count1.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_count2.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_count3.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_count4.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_count5.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_trace_addr.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_trace_buf0.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_trace_buf1.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_trace_buf2.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_trace_buf3.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_trace_buf4.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_misr0.attr.attr,
|
|
&coresight_attr_cfg_dbgbus_misr1.attr.attr,
|
|
&coresight_attr_ahb_dbg_cntl.attr.attr,
|
|
&coresight_attr_read_ahb_through_dbg.attr.attr,
|
|
&coresight_attr_dbg_lo_hi_gpio.attr.attr,
|
|
&coresight_attr_ext_trace_bus_cntl.attr.attr,
|
|
&coresight_attr_ext_vbif_dbg_cntl.attr.attr,
|
|
NULL,
|
|
};
|
|
|
|
static const struct attribute_group a5xx_coresight_group = {
|
|
.attrs = a5xx_coresight_attrs,
|
|
};
|
|
|
|
static const struct attribute_group *a5xx_coresight_groups[] = {
|
|
&a5xx_coresight_group,
|
|
NULL,
|
|
};
|
|
|
|
static struct adreno_coresight a5xx_coresight = {
|
|
.registers = a5xx_coresight_registers,
|
|
.count = ARRAY_SIZE(a5xx_coresight_registers),
|
|
.groups = a5xx_coresight_groups,
|
|
.read = kgsl_regread,
|
|
.write = kgsl_regwrite,
|
|
};
|
|
|
|
struct adreno_gpudev adreno_a5xx_gpudev = {
|
|
.reg_offsets = &a5xx_reg_offsets,
|
|
.int_bits = a5xx_int_bits,
|
|
.ft_perf_counters = a5xx_ft_perf_counters,
|
|
.ft_perf_counters_count = ARRAY_SIZE(a5xx_ft_perf_counters),
|
|
.coresight = {&a5xx_coresight},
|
|
.start = a5xx_start,
|
|
.snapshot = a5xx_snapshot,
|
|
.irq = &a5xx_irq,
|
|
.snapshot_data = &a5xx_snapshot_data,
|
|
.irq_trace = trace_kgsl_a5xx_irq_status,
|
|
.num_prio_levels = KGSL_PRIORITY_MAX_RB_LEVELS,
|
|
.platform_setup = a5xx_platform_setup,
|
|
.init = a5xx_init,
|
|
.remove = a5xx_remove,
|
|
.rb_start = a5xx_rb_start,
|
|
.microcode_read = a5xx_microcode_read,
|
|
.perfcounters = &a5xx_perfcounters,
|
|
.vbif_xin_halt_ctrl0_mask = A5XX_VBIF_XIN_HALT_CTRL0_MASK,
|
|
.is_sptp_idle = a5xx_is_sptp_idle,
|
|
.regulator_enable = a5xx_regulator_enable,
|
|
.regulator_disable = a5xx_regulator_disable,
|
|
.pwrlevel_change_settings = a5xx_pwrlevel_change_settings,
|
|
.read_throttling_counters = a5xx_read_throttling_counters,
|
|
.count_throttles = a5xx_count_throttles,
|
|
.enable_pwr_counters = a5xx_enable_pwr_counters,
|
|
.preemption_pre_ibsubmit = a5xx_preemption_pre_ibsubmit,
|
|
.preemption_yield_enable =
|
|
a5xx_preemption_yield_enable,
|
|
.preemption_post_ibsubmit =
|
|
a5xx_preemption_post_ibsubmit,
|
|
.preemption_init = a5xx_preemption_init,
|
|
.preemption_close = a5xx_preemption_close,
|
|
.preemption_schedule = a5xx_preemption_schedule,
|
|
.enable_64bit = a5xx_enable_64bit,
|
|
.clk_set_options = a5xx_clk_set_options,
|
|
.snapshot_preemption = a5xx_snapshot_preemption,
|
|
.zap_shader_unload = a5xx_zap_shader_unload,
|
|
};
|
|
|