You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
883 lines
24 KiB
883 lines
24 KiB
/* Copyright (c) 2010-2019, The Linux Foundation. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 and
|
|
* only version 2 as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
*/
|
|
|
|
#include <linux/export.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/hrtimer.h>
|
|
#include <linux/devfreq_cooling.h>
|
|
#include <linux/pm_opp.h>
|
|
|
|
#include "kgsl.h"
|
|
#include "kgsl_pwrscale.h"
|
|
#include "kgsl_device.h"
|
|
#include "kgsl_trace.h"
|
|
|
|
/**
|
|
* struct kgsl_midframe_info - midframe power stats sampling info
|
|
* @timer - midframe sampling timer
|
|
* @timer_check_ws - Updates powerstats on midframe expiry
|
|
* @device - pointer to kgsl_device
|
|
*/
|
|
static struct kgsl_midframe_info {
|
|
struct hrtimer timer;
|
|
struct work_struct timer_check_ws;
|
|
struct kgsl_device *device;
|
|
} *kgsl_midframe = NULL;
|
|
|
|
static void do_devfreq_suspend(struct work_struct *work);
|
|
static void do_devfreq_resume(struct work_struct *work);
|
|
static void do_devfreq_notify(struct work_struct *work);
|
|
|
|
/*
|
|
* These variables are used to keep the latest data
|
|
* returned by kgsl_devfreq_get_dev_status
|
|
*/
|
|
static struct xstats last_xstats;
|
|
static struct devfreq_dev_status last_status = { .private_data = &last_xstats };
|
|
|
|
/*
|
|
* kgsl_pwrscale_sleep - notify governor that device is going off
|
|
* @device: The device
|
|
*
|
|
* Called shortly after all pending work is completed.
|
|
*/
|
|
void kgsl_pwrscale_sleep(struct kgsl_device *device)
|
|
{
|
|
if (!device->pwrscale.enabled)
|
|
return;
|
|
device->pwrscale.on_time = 0;
|
|
|
|
/* to call devfreq_suspend_device() from a kernel thread */
|
|
queue_work(device->pwrscale.devfreq_wq,
|
|
&device->pwrscale.devfreq_suspend_ws);
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_sleep);
|
|
|
|
/*
|
|
* kgsl_pwrscale_wake - notify governor that device is going on
|
|
* @device: The device
|
|
*
|
|
* Called when the device is returning to an active state.
|
|
*/
|
|
void kgsl_pwrscale_wake(struct kgsl_device *device)
|
|
{
|
|
struct kgsl_power_stats stats;
|
|
struct kgsl_pwrscale *psc = &device->pwrscale;
|
|
|
|
if (!device->pwrscale.enabled)
|
|
return;
|
|
/* clear old stats before waking */
|
|
memset(&psc->accum_stats, 0, sizeof(psc->accum_stats));
|
|
memset(&last_xstats, 0, sizeof(last_xstats));
|
|
|
|
/* and any hw activity from waking up*/
|
|
device->ftbl->power_stats(device, &stats);
|
|
|
|
psc->time = ktime_get();
|
|
|
|
psc->next_governor_call = ktime_add_us(psc->time,
|
|
KGSL_GOVERNOR_CALL_INTERVAL);
|
|
|
|
/* to call devfreq_resume_device() from a kernel thread */
|
|
queue_work(psc->devfreq_wq, &psc->devfreq_resume_ws);
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_wake);
|
|
|
|
/*
|
|
* kgsl_pwrscale_busy - update pwrscale state for new work
|
|
* @device: The device
|
|
*
|
|
* Called when new work is submitted to the device.
|
|
* This function must be called with the device mutex locked.
|
|
*/
|
|
void kgsl_pwrscale_busy(struct kgsl_device *device)
|
|
{
|
|
if (!device->pwrscale.enabled)
|
|
return;
|
|
if (device->pwrscale.on_time == 0)
|
|
device->pwrscale.on_time = ktime_to_us(ktime_get());
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_busy);
|
|
|
|
/**
|
|
* kgsl_pwrscale_update_stats() - update device busy statistics
|
|
* @device: The device
|
|
*
|
|
* Read hardware busy counters and accumulate the results.
|
|
*/
|
|
void kgsl_pwrscale_update_stats(struct kgsl_device *device)
|
|
{
|
|
struct kgsl_pwrctrl *pwrctrl = &device->pwrctrl;
|
|
struct kgsl_pwrscale *psc = &device->pwrscale;
|
|
|
|
if (WARN_ON(!mutex_is_locked(&device->mutex)))
|
|
return;
|
|
|
|
if (!psc->enabled)
|
|
return;
|
|
|
|
if (device->state == KGSL_STATE_ACTIVE) {
|
|
struct kgsl_power_stats stats;
|
|
|
|
device->ftbl->power_stats(device, &stats);
|
|
device->pwrscale.accum_stats.busy_time += stats.busy_time;
|
|
device->pwrscale.accum_stats.ram_time += stats.ram_time;
|
|
device->pwrscale.accum_stats.ram_wait += stats.ram_wait;
|
|
pwrctrl->clock_times[pwrctrl->active_pwrlevel] +=
|
|
stats.busy_time;
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_update_stats);
|
|
|
|
/**
|
|
* kgsl_pwrscale_update() - update device busy statistics
|
|
* @device: The device
|
|
*
|
|
* If enough time has passed schedule the next call to devfreq
|
|
* get_dev_status.
|
|
*/
|
|
void kgsl_pwrscale_update(struct kgsl_device *device)
|
|
{
|
|
ktime_t t;
|
|
|
|
if (WARN_ON(!mutex_is_locked(&device->mutex)))
|
|
return;
|
|
|
|
if (!device->pwrscale.enabled)
|
|
return;
|
|
|
|
t = ktime_get();
|
|
if (ktime_compare(t, device->pwrscale.next_governor_call) < 0)
|
|
return;
|
|
|
|
device->pwrscale.next_governor_call = ktime_add_us(t,
|
|
KGSL_GOVERNOR_CALL_INTERVAL);
|
|
|
|
/* to call update_devfreq() from a kernel thread */
|
|
if (device->state != KGSL_STATE_SLUMBER)
|
|
queue_work(device->pwrscale.devfreq_wq,
|
|
&device->pwrscale.devfreq_notify_ws);
|
|
|
|
kgsl_pwrscale_midframe_timer_restart(device);
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_update);
|
|
|
|
void kgsl_pwrscale_midframe_timer_restart(struct kgsl_device *device)
|
|
{
|
|
if (kgsl_midframe) {
|
|
WARN_ON(!mutex_is_locked(&device->mutex));
|
|
|
|
/* If the timer is already running, stop it */
|
|
if (hrtimer_active(&kgsl_midframe->timer))
|
|
hrtimer_cancel(
|
|
&kgsl_midframe->timer);
|
|
|
|
hrtimer_start(&kgsl_midframe->timer,
|
|
ns_to_ktime(KGSL_GOVERNOR_CALL_INTERVAL
|
|
* NSEC_PER_USEC), HRTIMER_MODE_REL);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_midframe_timer_restart);
|
|
|
|
void kgsl_pwrscale_midframe_timer_cancel(struct kgsl_device *device)
|
|
{
|
|
if (kgsl_midframe) {
|
|
WARN_ON(!mutex_is_locked(&device->mutex));
|
|
hrtimer_cancel(&kgsl_midframe->timer);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_midframe_timer_cancel);
|
|
|
|
static void kgsl_pwrscale_midframe_timer_check(struct work_struct *work)
|
|
{
|
|
struct kgsl_device *device = kgsl_midframe->device;
|
|
|
|
mutex_lock(&device->mutex);
|
|
if (device->state == KGSL_STATE_ACTIVE)
|
|
kgsl_pwrscale_update(device);
|
|
mutex_unlock(&device->mutex);
|
|
}
|
|
|
|
static enum hrtimer_restart kgsl_pwrscale_midframe_timer(struct hrtimer *timer)
|
|
{
|
|
struct kgsl_device *device = kgsl_midframe->device;
|
|
|
|
queue_work(device->pwrscale.devfreq_wq,
|
|
&kgsl_midframe->timer_check_ws);
|
|
|
|
return HRTIMER_NORESTART;
|
|
}
|
|
|
|
/*
|
|
* kgsl_pwrscale_disable - temporarily disable the governor
|
|
* @device: The device
|
|
* @turbo: Indicates if pwrlevel should be forced to turbo
|
|
*
|
|
* Temporarily disable the governor, to prevent interference
|
|
* with profiling tools that expect a fixed clock frequency.
|
|
* This function must be called with the device mutex locked.
|
|
*/
|
|
void kgsl_pwrscale_disable(struct kgsl_device *device, bool turbo)
|
|
{
|
|
if (WARN_ON(!mutex_is_locked(&device->mutex)))
|
|
return;
|
|
|
|
if (device->pwrscale.devfreqptr)
|
|
queue_work(device->pwrscale.devfreq_wq,
|
|
&device->pwrscale.devfreq_suspend_ws);
|
|
device->pwrscale.enabled = false;
|
|
if (turbo)
|
|
kgsl_pwrctrl_pwrlevel_change(device, KGSL_PWRLEVEL_TURBO);
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_disable);
|
|
|
|
/*
|
|
* kgsl_pwrscale_enable - re-enable the governor
|
|
* @device: The device
|
|
*
|
|
* Reenable the governor after a kgsl_pwrscale_disable() call.
|
|
* This function must be called with the device mutex locked.
|
|
*/
|
|
void kgsl_pwrscale_enable(struct kgsl_device *device)
|
|
{
|
|
if (WARN_ON(!mutex_is_locked(&device->mutex)))
|
|
return;
|
|
|
|
if (device->pwrscale.devfreqptr) {
|
|
queue_work(device->pwrscale.devfreq_wq,
|
|
&device->pwrscale.devfreq_resume_ws);
|
|
device->pwrscale.enabled = true;
|
|
} else {
|
|
/*
|
|
* Don't enable it if devfreq is not set and let the device
|
|
* run at default level;
|
|
*/
|
|
kgsl_pwrctrl_pwrlevel_change(device,
|
|
device->pwrctrl.num_pwrlevels - 1);
|
|
device->pwrscale.enabled = false;
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_enable);
|
|
|
|
static int _thermal_adjust(struct kgsl_pwrctrl *pwr, int level)
|
|
{
|
|
if (level < pwr->active_pwrlevel)
|
|
return pwr->active_pwrlevel;
|
|
|
|
/*
|
|
* A lower frequency has been recommended! Stop thermal
|
|
* cycling (but keep the upper thermal limit) and switch to
|
|
* the lower frequency.
|
|
*/
|
|
pwr->thermal_cycle = CYCLE_ENABLE;
|
|
del_timer_sync(&pwr->thermal_timer);
|
|
return level;
|
|
}
|
|
|
|
#ifdef DEVFREQ_FLAG_WAKEUP_MAXFREQ
|
|
static inline bool _check_maxfreq(u32 flags)
|
|
{
|
|
return (flags & DEVFREQ_FLAG_WAKEUP_MAXFREQ);
|
|
}
|
|
#else
|
|
static inline bool _check_maxfreq(u32 flags)
|
|
{
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* kgsl_devfreq_target - devfreq_dev_profile.target callback
|
|
* @dev: see devfreq.h
|
|
* @freq: see devfreq.h
|
|
* @flags: see devfreq.h
|
|
*
|
|
* This function expects the device mutex to be unlocked.
|
|
*/
|
|
int kgsl_devfreq_target(struct device *dev, unsigned long *freq, u32 flags)
|
|
{
|
|
struct kgsl_device *device = dev_get_drvdata(dev);
|
|
struct kgsl_pwrctrl *pwr;
|
|
struct kgsl_pwrlevel *pwr_level;
|
|
int level;
|
|
unsigned int i;
|
|
unsigned long cur_freq, rec_freq;
|
|
|
|
if (device == NULL)
|
|
return -ENODEV;
|
|
if (freq == NULL)
|
|
return -EINVAL;
|
|
if (!device->pwrscale.enabled)
|
|
return 0;
|
|
|
|
pwr = &device->pwrctrl;
|
|
if (_check_maxfreq(flags)) {
|
|
/*
|
|
* The GPU is about to get suspended,
|
|
* but it needs to be at the max power level when waking up
|
|
*/
|
|
pwr->wakeup_maxpwrlevel = 1;
|
|
return 0;
|
|
}
|
|
|
|
rec_freq = *freq;
|
|
|
|
mutex_lock(&device->mutex);
|
|
cur_freq = kgsl_pwrctrl_active_freq(pwr);
|
|
level = pwr->active_pwrlevel;
|
|
pwr_level = &pwr->pwrlevels[level];
|
|
|
|
/* If the governor recommends a new frequency, update it here */
|
|
if (rec_freq != cur_freq) {
|
|
level = pwr->max_pwrlevel;
|
|
/*
|
|
* Array index of pwrlevels[] should be within the permitted
|
|
* power levels, i.e., from max_pwrlevel to min_pwrlevel.
|
|
*/
|
|
for (i = pwr->min_pwrlevel; (i >= pwr->max_pwrlevel
|
|
&& i <= pwr->min_pwrlevel); i--)
|
|
if (rec_freq <= pwr->pwrlevels[i].gpu_freq) {
|
|
if (pwr->thermal_cycle == CYCLE_ACTIVE)
|
|
level = _thermal_adjust(pwr, i);
|
|
else
|
|
level = i;
|
|
break;
|
|
}
|
|
if (level != pwr->active_pwrlevel)
|
|
kgsl_pwrctrl_pwrlevel_change(device, level);
|
|
}
|
|
|
|
*freq = kgsl_pwrctrl_active_freq(pwr);
|
|
|
|
mutex_unlock(&device->mutex);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(kgsl_devfreq_target);
|
|
|
|
/*
|
|
* kgsl_devfreq_get_dev_status - devfreq_dev_profile.get_dev_status callback
|
|
* @dev: see devfreq.h
|
|
* @freq: see devfreq.h
|
|
* @flags: see devfreq.h
|
|
*
|
|
* This function expects the device mutex to be unlocked.
|
|
*/
|
|
int kgsl_devfreq_get_dev_status(struct device *dev,
|
|
struct devfreq_dev_status *stat)
|
|
{
|
|
struct kgsl_device *device = dev_get_drvdata(dev);
|
|
struct kgsl_pwrctrl *pwrctrl;
|
|
struct kgsl_pwrscale *pwrscale;
|
|
ktime_t tmp1, tmp2;
|
|
|
|
if (device == NULL)
|
|
return -ENODEV;
|
|
if (stat == NULL)
|
|
return -EINVAL;
|
|
|
|
pwrscale = &device->pwrscale;
|
|
pwrctrl = &device->pwrctrl;
|
|
|
|
mutex_lock(&device->mutex);
|
|
|
|
tmp1 = ktime_get();
|
|
/*
|
|
* If the GPU clock is on grab the latest power counter
|
|
* values. Otherwise the most recent ACTIVE values will
|
|
* already be stored in accum_stats.
|
|
*/
|
|
kgsl_pwrscale_update_stats(device);
|
|
|
|
tmp2 = ktime_get();
|
|
stat->total_time = ktime_us_delta(tmp2, pwrscale->time);
|
|
pwrscale->time = tmp1;
|
|
|
|
stat->busy_time = pwrscale->accum_stats.busy_time;
|
|
|
|
stat->current_frequency = kgsl_pwrctrl_active_freq(&device->pwrctrl);
|
|
|
|
stat->private_data = &device->active_context_count;
|
|
|
|
/*
|
|
* keep the latest devfreq_dev_status values
|
|
* and vbif counters data
|
|
* to be (re)used by kgsl_busmon_get_dev_status()
|
|
*/
|
|
if (pwrctrl->bus_control) {
|
|
struct xstats *last_b =
|
|
(struct xstats *)last_status.private_data;
|
|
|
|
last_status.total_time = stat->total_time;
|
|
last_status.busy_time = stat->busy_time;
|
|
last_status.current_frequency = stat->current_frequency;
|
|
|
|
last_b->ram_time = device->pwrscale.accum_stats.ram_time;
|
|
last_b->ram_wait = device->pwrscale.accum_stats.ram_wait;
|
|
last_b->mod = device->pwrctrl.bus_mod;
|
|
}
|
|
|
|
kgsl_pwrctrl_busy_time(device, stat->total_time, stat->busy_time);
|
|
trace_kgsl_pwrstats(device, stat->total_time,
|
|
&pwrscale->accum_stats, device->active_context_count);
|
|
memset(&pwrscale->accum_stats, 0, sizeof(pwrscale->accum_stats));
|
|
|
|
mutex_unlock(&device->mutex);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(kgsl_devfreq_get_dev_status);
|
|
|
|
/*
|
|
* kgsl_devfreq_get_cur_freq - devfreq_dev_profile.get_cur_freq callback
|
|
* @dev: see devfreq.h
|
|
* @freq: see devfreq.h
|
|
* @flags: see devfreq.h
|
|
*
|
|
* This function expects the device mutex to be unlocked.
|
|
*/
|
|
int kgsl_devfreq_get_cur_freq(struct device *dev, unsigned long *freq)
|
|
{
|
|
struct kgsl_device *device = dev_get_drvdata(dev);
|
|
|
|
if (device == NULL)
|
|
return -ENODEV;
|
|
if (freq == NULL)
|
|
return -EINVAL;
|
|
|
|
mutex_lock(&device->mutex);
|
|
*freq = kgsl_pwrctrl_active_freq(&device->pwrctrl);
|
|
mutex_unlock(&device->mutex);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(kgsl_devfreq_get_cur_freq);
|
|
|
|
/*
|
|
* kgsl_busmon_get_dev_status - devfreq_dev_profile.get_dev_status callback
|
|
* @dev: see devfreq.h
|
|
* @freq: see devfreq.h
|
|
* @flags: see devfreq.h
|
|
*
|
|
* This function expects the device mutex to be unlocked.
|
|
*/
|
|
int kgsl_busmon_get_dev_status(struct device *dev,
|
|
struct devfreq_dev_status *stat)
|
|
{
|
|
struct xstats *b;
|
|
|
|
stat->total_time = last_status.total_time;
|
|
stat->busy_time = last_status.busy_time;
|
|
stat->current_frequency = last_status.current_frequency;
|
|
if (stat->private_data) {
|
|
struct xstats *last_b =
|
|
(struct xstats *)last_status.private_data;
|
|
b = (struct xstats *)stat->private_data;
|
|
b->ram_time = last_b->ram_time;
|
|
b->ram_wait = last_b->ram_wait;
|
|
b->mod = last_b->mod;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#ifdef DEVFREQ_FLAG_FAST_HINT
|
|
static inline bool _check_fast_hint(u32 flags)
|
|
{
|
|
return (flags & DEVFREQ_FLAG_FAST_HINT);
|
|
}
|
|
#else
|
|
static inline bool _check_fast_hint(u32 flags)
|
|
{
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
#ifdef DEVFREQ_FLAG_SLOW_HINT
|
|
static inline bool _check_slow_hint(u32 flags)
|
|
{
|
|
return (flags & DEVFREQ_FLAG_SLOW_HINT);
|
|
}
|
|
#else
|
|
static inline bool _check_slow_hint(u32 flags)
|
|
{
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* kgsl_busmon_target - devfreq_dev_profile.target callback
|
|
* @dev: see devfreq.h
|
|
* @freq: see devfreq.h
|
|
* @flags: see devfreq.h
|
|
*
|
|
* This function expects the device mutex to be unlocked.
|
|
*/
|
|
int kgsl_busmon_target(struct device *dev, unsigned long *freq, u32 flags)
|
|
{
|
|
struct kgsl_device *device = dev_get_drvdata(dev);
|
|
struct kgsl_pwrctrl *pwr;
|
|
struct kgsl_pwrlevel *pwr_level;
|
|
int level, b;
|
|
u32 bus_flag;
|
|
unsigned long ab_mbytes;
|
|
|
|
if (device == NULL)
|
|
return -ENODEV;
|
|
if (freq == NULL)
|
|
return -EINVAL;
|
|
if (!device->pwrscale.enabled)
|
|
return 0;
|
|
|
|
pwr = &device->pwrctrl;
|
|
|
|
if (!pwr->bus_control)
|
|
return 0;
|
|
|
|
mutex_lock(&device->mutex);
|
|
level = pwr->active_pwrlevel;
|
|
pwr_level = &pwr->pwrlevels[level];
|
|
bus_flag = device->pwrscale.bus_profile.flag;
|
|
device->pwrscale.bus_profile.flag = 0;
|
|
ab_mbytes = device->pwrscale.bus_profile.ab_mbytes;
|
|
|
|
/*
|
|
* Bus devfreq governor has calculated its recomendations
|
|
* when gpu was running with *freq frequency.
|
|
* If the gpu frequency is different now it's better to
|
|
* ignore the call
|
|
*/
|
|
if (pwr_level->gpu_freq != *freq) {
|
|
mutex_unlock(&device->mutex);
|
|
return 0;
|
|
}
|
|
|
|
b = pwr->bus_mod;
|
|
if (_check_fast_hint(bus_flag))
|
|
pwr->bus_mod++;
|
|
else if (_check_slow_hint(bus_flag))
|
|
pwr->bus_mod--;
|
|
|
|
/* trim calculated change to fit range */
|
|
if (pwr_level->bus_freq + pwr->bus_mod < pwr_level->bus_min)
|
|
pwr->bus_mod = -(pwr_level->bus_freq - pwr_level->bus_min);
|
|
else if (pwr_level->bus_freq + pwr->bus_mod > pwr_level->bus_max)
|
|
pwr->bus_mod = pwr_level->bus_max - pwr_level->bus_freq;
|
|
|
|
/* Update bus vote if AB or IB is modified */
|
|
if ((pwr->bus_mod != b) || (pwr->bus_ab_mbytes != ab_mbytes)) {
|
|
pwr->bus_percent_ab = device->pwrscale.bus_profile.percent_ab;
|
|
pwr->bus_ab_mbytes = ab_mbytes;
|
|
kgsl_pwrctrl_buslevel_update(device, true);
|
|
}
|
|
|
|
mutex_unlock(&device->mutex);
|
|
return 0;
|
|
}
|
|
|
|
int kgsl_busmon_get_cur_freq(struct device *dev, unsigned long *freq)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* opp_notify - Callback function registered to receive OPP events.
|
|
* @nb: The notifier block
|
|
* @type: The event type. Two OPP events are expected in this function:
|
|
* - OPP_EVENT_ENABLE: an GPU OPP is enabled. The in_opp parameter
|
|
* contains the OPP that is enabled
|
|
* - OPP_EVENT_DISALBE: an GPU OPP is disabled. The in_opp parameter
|
|
* contains the OPP that is disabled.
|
|
* @in_opp: the GPU OPP whose status is changed and triggered the event
|
|
*
|
|
* GPU OPP event callback function. The function subscribe GPU OPP status
|
|
* change and update thermal power level accordingly.
|
|
*/
|
|
|
|
static int opp_notify(struct notifier_block *nb,
|
|
unsigned long type, void *in_opp)
|
|
{
|
|
int level, min_level, max_level;
|
|
struct kgsl_pwrctrl *pwr = container_of(nb, struct kgsl_pwrctrl, nb);
|
|
struct kgsl_device *device = container_of(pwr,
|
|
struct kgsl_device, pwrctrl);
|
|
struct device *dev = &device->pdev->dev;
|
|
struct dev_pm_opp *opp;
|
|
unsigned long min_freq = 0, max_freq = pwr->pwrlevels[0].gpu_freq;
|
|
|
|
if (type != OPP_EVENT_ENABLE && type != OPP_EVENT_DISABLE)
|
|
return -EINVAL;
|
|
|
|
opp = dev_pm_opp_find_freq_floor(dev, &max_freq);
|
|
if (IS_ERR(opp))
|
|
return PTR_ERR(opp);
|
|
|
|
dev_pm_opp_put(opp);
|
|
|
|
opp = dev_pm_opp_find_freq_ceil(dev, &min_freq);
|
|
if (IS_ERR(opp))
|
|
min_freq = pwr->pwrlevels[pwr->min_pwrlevel].gpu_freq;
|
|
else
|
|
dev_pm_opp_put(opp);
|
|
|
|
mutex_lock(&device->mutex);
|
|
|
|
max_level = pwr->thermal_pwrlevel;
|
|
min_level = pwr->thermal_pwrlevel_floor;
|
|
|
|
/* Thermal limit cannot be lower than lowest non-zero operating freq */
|
|
for (level = 0; level < (pwr->num_pwrlevels - 1); level++) {
|
|
if (pwr->pwrlevels[level].gpu_freq == max_freq)
|
|
max_level = level;
|
|
if (pwr->pwrlevels[level].gpu_freq == min_freq)
|
|
min_level = level;
|
|
}
|
|
|
|
pwr->thermal_pwrlevel = max_level;
|
|
pwr->thermal_pwrlevel_floor = min_level;
|
|
|
|
/* Update the current level using the new limit */
|
|
kgsl_pwrctrl_pwrlevel_change(device, pwr->active_pwrlevel);
|
|
mutex_unlock(&device->mutex);
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/*
|
|
* kgsl_pwrscale_init - Initialize pwrscale.
|
|
* @dev: The device
|
|
* @governor: The initial governor to use.
|
|
*
|
|
* Initialize devfreq and any non-constant profile data.
|
|
*/
|
|
int kgsl_pwrscale_init(struct device *dev, const char *governor)
|
|
{
|
|
struct kgsl_device *device;
|
|
struct kgsl_pwrscale *pwrscale;
|
|
struct kgsl_pwrctrl *pwr;
|
|
struct devfreq *devfreq;
|
|
struct msm_adreno_extended_profile *gpu_profile;
|
|
struct devfreq_dev_profile *profile;
|
|
struct devfreq_msm_adreno_tz_data *data;
|
|
int i, out = 0;
|
|
int ret;
|
|
|
|
device = dev_get_drvdata(dev);
|
|
if (device == NULL)
|
|
return -ENODEV;
|
|
|
|
pwrscale = &device->pwrscale;
|
|
pwr = &device->pwrctrl;
|
|
gpu_profile = &pwrscale->gpu_profile;
|
|
profile = &pwrscale->gpu_profile.profile;
|
|
|
|
pwr->nb.notifier_call = opp_notify;
|
|
|
|
dev_pm_opp_register_notifier(dev, &pwr->nb);
|
|
|
|
profile->initial_freq =
|
|
pwr->pwrlevels[pwr->num_pwrlevels - 1].gpu_freq;
|
|
/* Let's start with 10 ms and tune in later */
|
|
profile->polling_ms = 10;
|
|
|
|
/* do not include the 'off' level or duplicate freq. levels */
|
|
for (i = 0; i < (pwr->num_pwrlevels - 1); i++)
|
|
pwrscale->freq_table[out++] = pwr->pwrlevels[i].gpu_freq;
|
|
|
|
/*
|
|
* Max_state is the number of valid power levels.
|
|
* The valid power levels range from 0 - (max_state - 1)
|
|
*/
|
|
profile->max_state = pwr->num_pwrlevels - 1;
|
|
/* link storage array to the devfreq profile pointer */
|
|
profile->freq_table = pwrscale->freq_table;
|
|
|
|
/* if there is only 1 freq, no point in running a governor */
|
|
if (profile->max_state == 1)
|
|
governor = "performance";
|
|
|
|
/* initialize msm-adreno-tz governor specific data here */
|
|
data = gpu_profile->private_data;
|
|
|
|
data->disable_busy_time_burst = of_property_read_bool(
|
|
device->pdev->dev.of_node, "qcom,disable-busy-time-burst");
|
|
|
|
if (pwrscale->ctxt_aware_enable) {
|
|
data->ctxt_aware_enable = pwrscale->ctxt_aware_enable;
|
|
data->bin.ctxt_aware_target_pwrlevel =
|
|
pwrscale->ctxt_aware_target_pwrlevel;
|
|
data->bin.ctxt_aware_busy_penalty =
|
|
pwrscale->ctxt_aware_busy_penalty;
|
|
}
|
|
|
|
if (of_property_read_bool(device->pdev->dev.of_node,
|
|
"qcom,enable-midframe-timer")) {
|
|
kgsl_midframe = kzalloc(
|
|
sizeof(struct kgsl_midframe_info), GFP_KERNEL);
|
|
if (kgsl_midframe) {
|
|
hrtimer_init(&kgsl_midframe->timer,
|
|
CLOCK_MONOTONIC, HRTIMER_MODE_REL);
|
|
kgsl_midframe->timer.function =
|
|
kgsl_pwrscale_midframe_timer;
|
|
kgsl_midframe->device = device;
|
|
} else
|
|
KGSL_PWR_ERR(device,
|
|
"Failed to enable-midframe-timer feature\n");
|
|
}
|
|
|
|
/*
|
|
* If there is a separate GX power rail, allow
|
|
* independent modification to its voltage through
|
|
* the bus bandwidth vote.
|
|
*/
|
|
if (pwr->bus_control) {
|
|
out = 0;
|
|
while (pwr->bus_ib[out] && out <= pwr->pwrlevels[0].bus_max) {
|
|
pwr->bus_ib[out] =
|
|
pwr->bus_ib[out] >> 20;
|
|
out++;
|
|
}
|
|
data->bus.num = out;
|
|
data->bus.ib = &pwr->bus_ib[0];
|
|
data->bus.index = &pwr->bus_index[0];
|
|
data->bus.width = pwr->bus_width;
|
|
} else
|
|
data->bus.num = 0;
|
|
|
|
devfreq = devfreq_add_device(dev, &pwrscale->gpu_profile.profile,
|
|
governor, pwrscale->gpu_profile.private_data);
|
|
if (IS_ERR(devfreq)) {
|
|
device->pwrscale.enabled = false;
|
|
return PTR_ERR(devfreq);
|
|
}
|
|
|
|
pwrscale->devfreqptr = devfreq;
|
|
pwrscale->cooling_dev = of_devfreq_cooling_register(
|
|
device->pdev->dev.of_node, devfreq);
|
|
if (IS_ERR(pwrscale->cooling_dev))
|
|
pwrscale->cooling_dev = NULL;
|
|
|
|
if (data->bus.num) {
|
|
pwrscale->bus_profile.profile.max_state
|
|
= pwr->num_pwrlevels - 1;
|
|
pwrscale->bus_profile.profile.freq_table
|
|
= pwrscale->freq_table;
|
|
|
|
pwrscale->bus_devfreq = devfreq_add_device(device->busmondev,
|
|
&pwrscale->bus_profile.profile, "gpubw_mon", NULL);
|
|
if (IS_ERR(pwrscale->bus_devfreq))
|
|
pwrscale->bus_devfreq = NULL;
|
|
}
|
|
|
|
ret = sysfs_create_link(&device->dev->kobj,
|
|
&devfreq->dev.kobj, "devfreq");
|
|
|
|
pwrscale->devfreq_wq = create_freezable_workqueue("kgsl_devfreq_wq");
|
|
INIT_WORK(&pwrscale->devfreq_suspend_ws, do_devfreq_suspend);
|
|
INIT_WORK(&pwrscale->devfreq_resume_ws, do_devfreq_resume);
|
|
INIT_WORK(&pwrscale->devfreq_notify_ws, do_devfreq_notify);
|
|
if (kgsl_midframe)
|
|
INIT_WORK(&kgsl_midframe->timer_check_ws,
|
|
kgsl_pwrscale_midframe_timer_check);
|
|
|
|
pwrscale->next_governor_call = ktime_add_us(ktime_get(),
|
|
KGSL_GOVERNOR_CALL_INTERVAL);
|
|
|
|
/* history tracking */
|
|
for (i = 0; i < KGSL_PWREVENT_MAX; i++) {
|
|
pwrscale->history[i].events = kcalloc(
|
|
pwrscale->history[i].size,
|
|
sizeof(struct kgsl_pwr_event), GFP_KERNEL);
|
|
pwrscale->history[i].type = i;
|
|
}
|
|
|
|
/* Add links to the devfreq sysfs nodes */
|
|
kgsl_gpu_sysfs_add_link(device->gpu_sysfs_kobj,
|
|
&pwrscale->devfreqptr->dev.kobj, "governor",
|
|
"gpu_governor");
|
|
kgsl_gpu_sysfs_add_link(device->gpu_sysfs_kobj,
|
|
&pwrscale->devfreqptr->dev.kobj,
|
|
"available_governors", "gpu_available_governor");
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_init);
|
|
|
|
/*
|
|
* kgsl_pwrscale_close - clean up pwrscale
|
|
* @device: the device
|
|
*
|
|
* This function should be called with the device mutex locked.
|
|
*/
|
|
void kgsl_pwrscale_close(struct kgsl_device *device)
|
|
{
|
|
int i;
|
|
struct kgsl_pwrscale *pwrscale;
|
|
struct kgsl_pwrctrl *pwr;
|
|
|
|
pwr = &device->pwrctrl;
|
|
pwrscale = &device->pwrscale;
|
|
if (!pwrscale->devfreqptr)
|
|
return;
|
|
if (pwrscale->cooling_dev)
|
|
devfreq_cooling_unregister(pwrscale->cooling_dev);
|
|
|
|
kgsl_pwrscale_midframe_timer_cancel(device);
|
|
flush_workqueue(pwrscale->devfreq_wq);
|
|
destroy_workqueue(pwrscale->devfreq_wq);
|
|
devfreq_remove_device(device->pwrscale.devfreqptr);
|
|
devfreq_remove_device(pwrscale->bus_devfreq);
|
|
kfree(kgsl_midframe);
|
|
kgsl_midframe = NULL;
|
|
device->pwrscale.bus_devfreq = NULL;
|
|
device->pwrscale.devfreqptr = NULL;
|
|
dev_pm_opp_unregister_notifier(&device->pdev->dev, &pwr->nb);
|
|
for (i = 0; i < KGSL_PWREVENT_MAX; i++)
|
|
kfree(pwrscale->history[i].events);
|
|
}
|
|
EXPORT_SYMBOL(kgsl_pwrscale_close);
|
|
|
|
static void do_devfreq_suspend(struct work_struct *work)
|
|
{
|
|
struct kgsl_pwrscale *pwrscale = container_of(work,
|
|
struct kgsl_pwrscale, devfreq_suspend_ws);
|
|
|
|
devfreq_suspend_device(pwrscale->devfreqptr);
|
|
devfreq_suspend_device(pwrscale->bus_devfreq);
|
|
}
|
|
|
|
static void do_devfreq_resume(struct work_struct *work)
|
|
{
|
|
struct kgsl_pwrscale *pwrscale = container_of(work,
|
|
struct kgsl_pwrscale, devfreq_resume_ws);
|
|
|
|
devfreq_resume_device(pwrscale->devfreqptr);
|
|
devfreq_resume_device(pwrscale->bus_devfreq);
|
|
}
|
|
|
|
static void do_devfreq_notify(struct work_struct *work)
|
|
{
|
|
struct kgsl_pwrscale *pwrscale = container_of(work,
|
|
struct kgsl_pwrscale, devfreq_notify_ws);
|
|
struct devfreq *devfreq = pwrscale->devfreqptr;
|
|
struct devfreq *bus_devfreq = pwrscale->bus_devfreq;
|
|
|
|
mutex_lock(&devfreq->lock);
|
|
update_devfreq(devfreq);
|
|
mutex_unlock(&devfreq->lock);
|
|
|
|
if (bus_devfreq) {
|
|
mutex_lock(&bus_devfreq->lock);
|
|
update_devfreq(bus_devfreq);
|
|
mutex_unlock(&bus_devfreq->lock);
|
|
}
|
|
}
|
|
|