2019-04-02 13:32:01 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2011-07-01 20:12:45 +00:00
|
|
|
/*
|
|
|
|
* drivers/base/power/domain.c - Common code related to device power domains.
|
|
|
|
*
|
|
|
|
* Copyright (C) 2011 Rafael J. Wysocki <rjw@sisk.pl>, Renesas Electronics Corp.
|
|
|
|
*/
|
2019-03-04 17:14:38 +00:00
|
|
|
#define pr_fmt(fmt) "PM: " fmt
|
|
|
|
|
2015-06-26 09:14:14 +00:00
|
|
|
#include <linux/delay.h>
|
2024-10-30 12:55:10 +00:00
|
|
|
#include <linux/idr.h>
|
2011-07-01 20:12:45 +00:00
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/io.h>
|
2014-09-19 18:27:36 +00:00
|
|
|
#include <linux/platform_device.h>
|
2018-04-05 10:23:34 +00:00
|
|
|
#include <linux/pm_opp.h>
|
2011-07-01 20:12:45 +00:00
|
|
|
#include <linux/pm_runtime.h>
|
|
|
|
#include <linux/pm_domain.h>
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
#include <linux/pm_qos.h>
|
2014-12-01 11:50:21 +00:00
|
|
|
#include <linux/pm_clock.h>
|
2011-07-01 20:12:45 +00:00
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/err.h>
|
2011-07-11 22:39:29 +00:00
|
|
|
#include <linux/sched.h>
|
|
|
|
#include <linux/suspend.h>
|
2011-11-27 12:11:36 +00:00
|
|
|
#include <linux/export.h>
|
2019-03-27 14:35:46 +00:00
|
|
|
#include <linux/cpu.h>
|
2020-12-08 19:19:55 +00:00
|
|
|
#include <linux/debugfs.h>
|
2011-11-27 12:11:36 +00:00
|
|
|
|
2024-10-30 12:55:10 +00:00
|
|
|
/* Provides a unique ID for each genpd device */
|
|
|
|
static DEFINE_IDA(genpd_ida);
|
|
|
|
|
2015-06-26 09:14:14 +00:00
|
|
|
#define GENPD_RETRY_MAX_MS 250 /* Approximate */
|
|
|
|
|
2011-11-27 12:11:36 +00:00
|
|
|
#define GENPD_DEV_CALLBACK(genpd, type, callback, dev) \
|
|
|
|
({ \
|
|
|
|
type (*__routine)(struct device *__d); \
|
|
|
|
type __ret = (type)0; \
|
|
|
|
\
|
|
|
|
__routine = genpd->dev_ops.callback; \
|
|
|
|
if (__routine) { \
|
|
|
|
__ret = __routine(dev); \
|
|
|
|
} \
|
|
|
|
__ret; \
|
|
|
|
})
|
2011-07-01 20:12:45 +00:00
|
|
|
|
2011-07-13 10:31:52 +00:00
|
|
|
static LIST_HEAD(gpd_list);
|
|
|
|
static DEFINE_MUTEX(gpd_list_lock);
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
struct genpd_lock_ops {
|
|
|
|
void (*lock)(struct generic_pm_domain *genpd);
|
|
|
|
void (*lock_nested)(struct generic_pm_domain *genpd, int depth);
|
|
|
|
int (*lock_interruptible)(struct generic_pm_domain *genpd);
|
|
|
|
void (*unlock)(struct generic_pm_domain *genpd);
|
|
|
|
};
|
|
|
|
|
|
|
|
static void genpd_lock_mtx(struct generic_pm_domain *genpd)
|
|
|
|
{
|
|
|
|
mutex_lock(&genpd->mlock);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_lock_nested_mtx(struct generic_pm_domain *genpd,
|
|
|
|
int depth)
|
|
|
|
{
|
|
|
|
mutex_lock_nested(&genpd->mlock, depth);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int genpd_lock_interruptible_mtx(struct generic_pm_domain *genpd)
|
|
|
|
{
|
|
|
|
return mutex_lock_interruptible(&genpd->mlock);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_unlock_mtx(struct generic_pm_domain *genpd)
|
|
|
|
{
|
|
|
|
return mutex_unlock(&genpd->mlock);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct genpd_lock_ops genpd_mtx_ops = {
|
|
|
|
.lock = genpd_lock_mtx,
|
|
|
|
.lock_nested = genpd_lock_nested_mtx,
|
|
|
|
.lock_interruptible = genpd_lock_interruptible_mtx,
|
|
|
|
.unlock = genpd_unlock_mtx,
|
|
|
|
};
|
|
|
|
|
2016-10-14 17:47:55 +00:00
|
|
|
static void genpd_lock_spin(struct generic_pm_domain *genpd)
|
|
|
|
__acquires(&genpd->slock)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&genpd->slock, flags);
|
|
|
|
genpd->lock_flags = flags;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_lock_nested_spin(struct generic_pm_domain *genpd,
|
|
|
|
int depth)
|
|
|
|
__acquires(&genpd->slock)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave_nested(&genpd->slock, flags, depth);
|
|
|
|
genpd->lock_flags = flags;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int genpd_lock_interruptible_spin(struct generic_pm_domain *genpd)
|
|
|
|
__acquires(&genpd->slock)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&genpd->slock, flags);
|
|
|
|
genpd->lock_flags = flags;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_unlock_spin(struct generic_pm_domain *genpd)
|
|
|
|
__releases(&genpd->slock)
|
|
|
|
{
|
|
|
|
spin_unlock_irqrestore(&genpd->slock, genpd->lock_flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct genpd_lock_ops genpd_spin_ops = {
|
|
|
|
.lock = genpd_lock_spin,
|
|
|
|
.lock_nested = genpd_lock_nested_spin,
|
|
|
|
.lock_interruptible = genpd_lock_interruptible_spin,
|
|
|
|
.unlock = genpd_unlock_spin,
|
|
|
|
};
|
|
|
|
|
2024-05-27 14:25:51 +00:00
|
|
|
static void genpd_lock_raw_spin(struct generic_pm_domain *genpd)
|
|
|
|
__acquires(&genpd->raw_slock)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
raw_spin_lock_irqsave(&genpd->raw_slock, flags);
|
|
|
|
genpd->raw_lock_flags = flags;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_lock_nested_raw_spin(struct generic_pm_domain *genpd,
|
|
|
|
int depth)
|
|
|
|
__acquires(&genpd->raw_slock)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
raw_spin_lock_irqsave_nested(&genpd->raw_slock, flags, depth);
|
|
|
|
genpd->raw_lock_flags = flags;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int genpd_lock_interruptible_raw_spin(struct generic_pm_domain *genpd)
|
|
|
|
__acquires(&genpd->raw_slock)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
raw_spin_lock_irqsave(&genpd->raw_slock, flags);
|
|
|
|
genpd->raw_lock_flags = flags;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_unlock_raw_spin(struct generic_pm_domain *genpd)
|
|
|
|
__releases(&genpd->raw_slock)
|
|
|
|
{
|
|
|
|
raw_spin_unlock_irqrestore(&genpd->raw_slock, genpd->raw_lock_flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct genpd_lock_ops genpd_raw_spin_ops = {
|
|
|
|
.lock = genpd_lock_raw_spin,
|
|
|
|
.lock_nested = genpd_lock_nested_raw_spin,
|
|
|
|
.lock_interruptible = genpd_lock_interruptible_raw_spin,
|
|
|
|
.unlock = genpd_unlock_raw_spin,
|
|
|
|
};
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
#define genpd_lock(p) p->lock_ops->lock(p)
|
|
|
|
#define genpd_lock_nested(p, d) p->lock_ops->lock_nested(p, d)
|
|
|
|
#define genpd_lock_interruptible(p) p->lock_ops->lock_interruptible(p)
|
|
|
|
#define genpd_unlock(p) p->lock_ops->unlock(p)
|
|
|
|
|
2020-09-24 11:04:47 +00:00
|
|
|
#define genpd_status_on(genpd) (genpd->status == GENPD_STATE_ON)
|
2016-10-14 17:47:55 +00:00
|
|
|
#define genpd_is_irq_safe(genpd) (genpd->flags & GENPD_FLAG_IRQ_SAFE)
|
2017-03-20 10:19:21 +00:00
|
|
|
#define genpd_is_always_on(genpd) (genpd->flags & GENPD_FLAG_ALWAYS_ON)
|
2017-11-07 12:48:11 +00:00
|
|
|
#define genpd_is_active_wakeup(genpd) (genpd->flags & GENPD_FLAG_ACTIVE_WAKEUP)
|
2019-03-27 14:35:46 +00:00
|
|
|
#define genpd_is_cpu_domain(genpd) (genpd->flags & GENPD_FLAG_CPU_DOMAIN)
|
2019-04-30 15:06:11 +00:00
|
|
|
#define genpd_is_rpm_always_on(genpd) (genpd->flags & GENPD_FLAG_RPM_ALWAYS_ON)
|
2023-08-25 11:26:32 +00:00
|
|
|
#define genpd_is_opp_table_fw(genpd) (genpd->flags & GENPD_FLAG_OPP_TABLE_FW)
|
2024-10-30 12:55:10 +00:00
|
|
|
#define genpd_is_dev_name_fw(genpd) (genpd->flags & GENPD_FLAG_DEV_NAME_FW)
|
2016-10-14 17:47:55 +00:00
|
|
|
|
2022-05-11 14:56:54 +00:00
|
|
|
static inline bool irq_safe_dev_in_sleep_domain(struct device *dev,
|
2017-06-12 15:17:41 +00:00
|
|
|
const struct generic_pm_domain *genpd)
|
2016-10-14 17:47:55 +00:00
|
|
|
{
|
|
|
|
bool ret;
|
|
|
|
|
|
|
|
ret = pm_runtime_is_irq_safe(dev) && !genpd_is_irq_safe(genpd);
|
|
|
|
|
2017-03-20 10:19:23 +00:00
|
|
|
/*
|
2022-05-11 14:56:54 +00:00
|
|
|
* Warn once if an IRQ safe device is attached to a domain, which
|
|
|
|
* callbacks are allowed to sleep. This indicates a suboptimal
|
|
|
|
* configuration for PM, but it doesn't matter for an always on domain.
|
2017-03-20 10:19:23 +00:00
|
|
|
*/
|
2022-05-11 14:56:55 +00:00
|
|
|
if (genpd_is_always_on(genpd) || genpd_is_rpm_always_on(genpd))
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
if (ret)
|
2016-10-14 17:47:55 +00:00
|
|
|
dev_warn_once(dev, "PM domain %s will not be powered off\n",
|
2024-10-30 12:55:10 +00:00
|
|
|
dev_name(&genpd->dev));
|
2016-10-14 17:47:55 +00:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2019-08-29 14:48:05 +00:00
|
|
|
static int genpd_runtime_suspend(struct device *dev);
|
|
|
|
|
2015-03-20 17:20:33 +00:00
|
|
|
/*
|
|
|
|
* Get the generic PM domain for a particular struct device.
|
|
|
|
* This validates the struct device pointer, the PM domain pointer,
|
|
|
|
* and checks that the PM domain pointer is a real generic PM domain.
|
|
|
|
* Any failure results in NULL being returned.
|
|
|
|
*/
|
2019-08-29 14:48:05 +00:00
|
|
|
static struct generic_pm_domain *dev_to_genpd_safe(struct device *dev)
|
2015-03-20 17:20:33 +00:00
|
|
|
{
|
|
|
|
if (IS_ERR_OR_NULL(dev) || IS_ERR_OR_NULL(dev->pm_domain))
|
|
|
|
return NULL;
|
|
|
|
|
2019-08-29 14:48:05 +00:00
|
|
|
/* A genpd's always have its ->runtime_suspend() callback assigned. */
|
|
|
|
if (dev->pm_domain->ops.runtime_suspend == genpd_runtime_suspend)
|
|
|
|
return pd_to_genpd(dev->pm_domain);
|
2015-03-20 17:20:33 +00:00
|
|
|
|
2019-08-29 14:48:05 +00:00
|
|
|
return NULL;
|
2015-03-20 17:20:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This should only be used where we are certain that the pm_domain
|
|
|
|
* attached to the device is a genpd domain.
|
|
|
|
*/
|
|
|
|
static struct generic_pm_domain *dev_to_genpd(struct device *dev)
|
2011-07-01 20:13:10 +00:00
|
|
|
{
|
|
|
|
if (IS_ERR_OR_NULL(dev->pm_domain))
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
2011-07-01 20:13:19 +00:00
|
|
|
return pd_to_genpd(dev->pm_domain);
|
2011-07-01 20:13:10 +00:00
|
|
|
}
|
2011-07-01 20:12:45 +00:00
|
|
|
|
2024-04-09 10:23:58 +00:00
|
|
|
struct device *dev_to_genpd_dev(struct device *dev)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = dev_to_genpd(dev);
|
|
|
|
|
|
|
|
if (IS_ERR(genpd))
|
|
|
|
return ERR_CAST(genpd);
|
|
|
|
|
|
|
|
return &genpd->dev;
|
|
|
|
}
|
|
|
|
|
2017-06-12 15:17:41 +00:00
|
|
|
static int genpd_stop_dev(const struct generic_pm_domain *genpd,
|
|
|
|
struct device *dev)
|
2011-11-27 12:11:36 +00:00
|
|
|
{
|
2015-10-15 15:02:19 +00:00
|
|
|
return GENPD_DEV_CALLBACK(genpd, int, stop, dev);
|
2011-11-27 12:11:36 +00:00
|
|
|
}
|
|
|
|
|
2017-06-12 15:17:41 +00:00
|
|
|
static int genpd_start_dev(const struct generic_pm_domain *genpd,
|
|
|
|
struct device *dev)
|
2011-11-27 12:11:36 +00:00
|
|
|
{
|
2015-10-15 15:02:19 +00:00
|
|
|
return GENPD_DEV_CALLBACK(genpd, int, start, dev);
|
2011-11-27 12:11:36 +00:00
|
|
|
}
|
|
|
|
|
2011-08-08 21:43:04 +00:00
|
|
|
static bool genpd_sd_counter_dec(struct generic_pm_domain *genpd)
|
2011-07-01 20:12:45 +00:00
|
|
|
{
|
2011-08-08 21:43:04 +00:00
|
|
|
bool ret = false;
|
|
|
|
|
|
|
|
if (!WARN_ON(atomic_read(&genpd->sd_count) == 0))
|
|
|
|
ret = !!atomic_dec_and_test(&genpd->sd_count);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_sd_counter_inc(struct generic_pm_domain *genpd)
|
|
|
|
{
|
|
|
|
atomic_inc(&genpd->sd_count);
|
2014-03-17 17:06:10 +00:00
|
|
|
smp_mb__after_atomic();
|
2011-07-01 20:12:45 +00:00
|
|
|
}
|
|
|
|
|
2017-07-14 17:10:15 +00:00
|
|
|
#ifdef CONFIG_DEBUG_FS
|
2020-12-08 19:19:55 +00:00
|
|
|
static struct dentry *genpd_debugfs_dir;
|
|
|
|
|
|
|
|
static void genpd_debug_add(struct generic_pm_domain *genpd);
|
|
|
|
|
|
|
|
static void genpd_debug_remove(struct generic_pm_domain *genpd)
|
|
|
|
{
|
2022-07-05 17:16:49 +00:00
|
|
|
if (!genpd_debugfs_dir)
|
|
|
|
return;
|
|
|
|
|
2024-10-30 12:55:10 +00:00
|
|
|
debugfs_lookup_and_remove(dev_name(&genpd->dev), genpd_debugfs_dir);
|
2020-12-08 19:19:55 +00:00
|
|
|
}
|
|
|
|
|
2017-07-14 17:10:15 +00:00
|
|
|
static void genpd_update_accounting(struct generic_pm_domain *genpd)
|
|
|
|
{
|
2022-04-19 17:29:16 +00:00
|
|
|
u64 delta, now;
|
2017-07-14 17:10:15 +00:00
|
|
|
|
2022-04-19 17:29:16 +00:00
|
|
|
now = ktime_get_mono_fast_ns();
|
|
|
|
if (now <= genpd->accounting_time)
|
|
|
|
return;
|
|
|
|
|
|
|
|
delta = now - genpd->accounting_time;
|
2017-07-14 17:10:15 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If genpd->status is active, it means we are just
|
|
|
|
* out of off and so update the idle time and vice
|
|
|
|
* versa.
|
|
|
|
*/
|
2022-04-19 17:29:16 +00:00
|
|
|
if (genpd->status == GENPD_STATE_ON)
|
|
|
|
genpd->states[genpd->state_idx].idle_time += delta;
|
|
|
|
else
|
|
|
|
genpd->on_time += delta;
|
2017-07-14 17:10:15 +00:00
|
|
|
|
|
|
|
genpd->accounting_time = now;
|
|
|
|
}
|
|
|
|
#else
|
2020-12-08 19:19:55 +00:00
|
|
|
static inline void genpd_debug_add(struct generic_pm_domain *genpd) {}
|
|
|
|
static inline void genpd_debug_remove(struct generic_pm_domain *genpd) {}
|
2017-07-14 17:10:15 +00:00
|
|
|
static inline void genpd_update_accounting(struct generic_pm_domain *genpd) {}
|
|
|
|
#endif
|
|
|
|
|
2018-10-31 09:26:54 +00:00
|
|
|
static int _genpd_reeval_performance_state(struct generic_pm_domain *genpd,
|
|
|
|
unsigned int state)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain_data *pd_data;
|
|
|
|
struct pm_domain_data *pdd;
|
2018-11-02 09:10:19 +00:00
|
|
|
struct gpd_link *link;
|
2018-10-31 09:26:54 +00:00
|
|
|
|
|
|
|
/* New requested state is same as Max requested state */
|
|
|
|
if (state == genpd->performance_state)
|
|
|
|
return state;
|
|
|
|
|
|
|
|
/* New requested state is higher than Max requested state */
|
|
|
|
if (state > genpd->performance_state)
|
|
|
|
return state;
|
|
|
|
|
|
|
|
/* Traverse all devices within the domain */
|
|
|
|
list_for_each_entry(pdd, &genpd->dev_list, list_node) {
|
|
|
|
pd_data = to_gpd_data(pdd);
|
|
|
|
|
|
|
|
if (pd_data->performance_state > state)
|
|
|
|
state = pd_data->performance_state;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2018-11-02 09:10:19 +00:00
|
|
|
* Traverse all sub-domains within the domain. This can be
|
|
|
|
* done without any additional locking as the link->performance_state
|
2020-07-08 23:32:13 +00:00
|
|
|
* field is protected by the parent genpd->lock, which is already taken.
|
2018-11-02 09:10:19 +00:00
|
|
|
*
|
|
|
|
* Also note that link->performance_state (subdomain's performance state
|
2020-07-08 23:32:13 +00:00
|
|
|
* requirement to parent domain) is different from
|
|
|
|
* link->child->performance_state (current performance state requirement
|
2018-11-02 09:10:19 +00:00
|
|
|
* of the devices/sub-domains of the subdomain) and so can have a
|
|
|
|
* different value.
|
|
|
|
*
|
|
|
|
* Note that we also take vote from powered-off sub-domains into account
|
|
|
|
* as the same is done for devices right now.
|
2018-10-31 09:26:54 +00:00
|
|
|
*/
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry(link, &genpd->parent_links, parent_node) {
|
2018-11-02 09:10:19 +00:00
|
|
|
if (link->performance_state > state)
|
|
|
|
state = link->performance_state;
|
|
|
|
}
|
|
|
|
|
2018-10-31 09:26:54 +00:00
|
|
|
return state;
|
|
|
|
}
|
|
|
|
|
2021-01-20 21:12:30 +00:00
|
|
|
static int genpd_xlate_performance_state(struct generic_pm_domain *genpd,
|
|
|
|
struct generic_pm_domain *parent,
|
|
|
|
unsigned int pstate)
|
|
|
|
{
|
|
|
|
if (!parent->set_performance_state)
|
|
|
|
return pstate;
|
|
|
|
|
|
|
|
return dev_pm_opp_xlate_performance_state(genpd->opp_table,
|
|
|
|
parent->opp_table,
|
|
|
|
pstate);
|
|
|
|
}
|
|
|
|
|
2018-10-31 09:26:54 +00:00
|
|
|
static int _genpd_set_performance_state(struct generic_pm_domain *genpd,
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
unsigned int state, int depth);
|
|
|
|
|
|
|
|
static void _genpd_rollback_parent_state(struct gpd_link *link, int depth)
|
2018-10-31 09:26:54 +00:00
|
|
|
{
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
struct generic_pm_domain *parent = link->parent;
|
|
|
|
int parent_state;
|
2018-10-31 09:26:54 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
genpd_lock_nested(parent, depth + 1);
|
2018-10-31 09:26:54 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
parent_state = link->prev_performance_state;
|
|
|
|
link->performance_state = parent_state;
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
parent_state = _genpd_reeval_performance_state(parent, parent_state);
|
|
|
|
if (_genpd_set_performance_state(parent, parent_state, depth + 1)) {
|
|
|
|
pr_err("%s: Failed to roll back to %d performance state\n",
|
|
|
|
parent->name, parent_state);
|
|
|
|
}
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
genpd_unlock(parent);
|
|
|
|
}
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
static int _genpd_set_parent_state(struct generic_pm_domain *genpd,
|
|
|
|
struct gpd_link *link,
|
|
|
|
unsigned int state, int depth)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *parent = link->parent;
|
|
|
|
int parent_state, ret;
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
/* Find parent's performance state */
|
|
|
|
ret = genpd_xlate_performance_state(genpd, parent, state);
|
|
|
|
if (unlikely(ret < 0))
|
|
|
|
return ret;
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
parent_state = ret;
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
genpd_lock_nested(parent, depth + 1);
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
link->prev_performance_state = link->performance_state;
|
|
|
|
link->performance_state = parent_state;
|
2018-10-31 09:26:54 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
parent_state = _genpd_reeval_performance_state(parent, parent_state);
|
|
|
|
ret = _genpd_set_performance_state(parent, parent_state, depth + 1);
|
|
|
|
if (ret)
|
|
|
|
link->performance_state = link->prev_performance_state;
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
genpd_unlock(parent);
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int _genpd_set_performance_state(struct generic_pm_domain *genpd,
|
|
|
|
unsigned int state, int depth)
|
|
|
|
{
|
|
|
|
struct gpd_link *link = NULL;
|
|
|
|
int ret;
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
if (state == genpd->performance_state)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* When scaling up, propagate to parents first in normal order */
|
|
|
|
if (state > genpd->performance_state) {
|
|
|
|
list_for_each_entry(link, &genpd->child_links, child_node) {
|
|
|
|
ret = _genpd_set_parent_state(genpd, link, state, depth);
|
|
|
|
if (ret)
|
|
|
|
goto rollback_parents_up;
|
|
|
|
}
|
|
|
|
}
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
if (genpd->set_performance_state) {
|
|
|
|
ret = genpd->set_performance_state(genpd, state);
|
|
|
|
if (ret) {
|
|
|
|
if (link)
|
|
|
|
goto rollback_parents_up;
|
|
|
|
return ret;
|
2018-11-02 09:10:19 +00:00
|
|
|
}
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
}
|
2018-11-02 09:10:19 +00:00
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
/* When scaling down, propagate to parents last in reverse order */
|
|
|
|
if (state < genpd->performance_state) {
|
|
|
|
list_for_each_entry_reverse(link, &genpd->child_links, child_node) {
|
|
|
|
ret = _genpd_set_parent_state(genpd, link, state, depth);
|
|
|
|
if (ret)
|
|
|
|
goto rollback_parents_down;
|
|
|
|
}
|
2018-11-02 09:10:19 +00:00
|
|
|
}
|
|
|
|
|
pmdomain: core: Scale down parent/child performance states in reverse order
Power domains might have parent domains assigned that are automatically
managed by the PM domain core. In particular, parent domains are
automatically powered on/off and setting performance states on child
domains are propagated to parent domains (e.g. using an OPP table from the
device tree).
Currently the parent performance state is always adjusted before the
performance state of the child domain, which is a problem for some cases
when scaling down the performance state. More exactly, it may lead to that
the parent domain could run in a lower performance state, than what is
required by the child domain.
To fix the behaviour, let's differentiate between scaling up/down and
adjust the order of operations:
- When scaling up, parent domains should be adjusted before the child
domain. In case of an error, the rollback happens in reverse order.
- When scaling down, parent domains should be adjusted after the child
domain, in reverse order, just as if we would rollback scaling up.
In case of an error, the rollback happens in normal order (just as
if we would normally scale up).
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Link: https://lore.kernel.org/r/20240103-genpd-perf-order-v2-1-eeecfc55624b@gerhold.net
Tested-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-01-03 21:10:05 +00:00
|
|
|
genpd->performance_state = state;
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
rollback_parents_up:
|
|
|
|
list_for_each_entry_continue_reverse(link, &genpd->child_links, child_node)
|
|
|
|
_genpd_rollback_parent_state(link, depth);
|
|
|
|
return ret;
|
|
|
|
rollback_parents_down:
|
|
|
|
list_for_each_entry_continue(link, &genpd->child_links, child_node)
|
|
|
|
_genpd_rollback_parent_state(link, depth);
|
2018-11-02 09:10:19 +00:00
|
|
|
return ret;
|
2018-10-31 09:26:54 +00:00
|
|
|
}
|
|
|
|
|
2021-06-03 09:34:35 +00:00
|
|
|
static int genpd_set_performance_state(struct device *dev, unsigned int state)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = dev_to_genpd(dev);
|
|
|
|
struct generic_pm_domain_data *gpd_data = dev_gpd_data(dev);
|
|
|
|
unsigned int prev_state;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
prev_state = gpd_data->performance_state;
|
2021-06-03 09:34:36 +00:00
|
|
|
if (prev_state == state)
|
|
|
|
return 0;
|
|
|
|
|
2021-06-03 09:34:35 +00:00
|
|
|
gpd_data->performance_state = state;
|
|
|
|
state = _genpd_reeval_performance_state(genpd, state);
|
|
|
|
|
|
|
|
ret = _genpd_set_performance_state(genpd, state, 0);
|
|
|
|
if (ret)
|
|
|
|
gpd_data->performance_state = prev_state;
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2021-06-03 09:34:37 +00:00
|
|
|
static int genpd_drop_performance_state(struct device *dev)
|
|
|
|
{
|
|
|
|
unsigned int prev_state = dev_gpd_data(dev)->performance_state;
|
|
|
|
|
|
|
|
if (!genpd_set_performance_state(dev, 0))
|
|
|
|
return prev_state;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_restore_performance_state(struct device *dev,
|
|
|
|
unsigned int state)
|
|
|
|
{
|
|
|
|
if (state)
|
|
|
|
genpd_set_performance_state(dev, state);
|
|
|
|
}
|
|
|
|
|
2023-09-25 13:17:08 +00:00
|
|
|
static int genpd_dev_pm_set_performance_state(struct device *dev,
|
|
|
|
unsigned int state)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = dev_to_genpd(dev);
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
genpd_lock(genpd);
|
|
|
|
if (pm_runtime_suspended(dev)) {
|
|
|
|
dev_gpd_data(dev)->rpm_pstate = state;
|
|
|
|
} else {
|
|
|
|
ret = genpd_set_performance_state(dev, state);
|
|
|
|
if (!ret)
|
|
|
|
dev_gpd_data(dev)->rpm_pstate = 0;
|
|
|
|
}
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2017-10-12 09:37:23 +00:00
|
|
|
/**
|
|
|
|
* dev_pm_genpd_set_performance_state- Set performance state of device's power
|
|
|
|
* domain.
|
|
|
|
*
|
|
|
|
* @dev: Device for which the performance-state needs to be set.
|
|
|
|
* @state: Target performance state of the device. This can be set as 0 when the
|
|
|
|
* device doesn't have any performance state constraints left (And so
|
|
|
|
* the device wouldn't participate anymore to find the target
|
|
|
|
* performance state of the genpd).
|
|
|
|
*
|
|
|
|
* It is assumed that the users guarantee that the genpd wouldn't be detached
|
|
|
|
* while this routine is getting called.
|
|
|
|
*
|
|
|
|
* Returns 0 on success and negative error values on failures.
|
|
|
|
*/
|
|
|
|
int dev_pm_genpd_set_performance_state(struct device *dev, unsigned int state)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
|
|
|
|
2019-08-29 14:48:15 +00:00
|
|
|
genpd = dev_to_genpd_safe(dev);
|
|
|
|
if (!genpd)
|
2017-10-12 09:37:23 +00:00
|
|
|
return -ENODEV;
|
|
|
|
|
2019-04-16 16:23:05 +00:00
|
|
|
if (WARN_ON(!dev->power.subsys_data ||
|
|
|
|
!dev->power.subsys_data->domain_data))
|
2017-10-12 09:37:23 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
2023-09-25 13:17:08 +00:00
|
|
|
return genpd_dev_pm_set_performance_state(dev, state);
|
2017-10-12 09:37:23 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_set_performance_state);
|
|
|
|
|
2021-01-20 15:50:41 +00:00
|
|
|
/**
|
|
|
|
* dev_pm_genpd_set_next_wakeup - Notify PM framework of an impending wakeup.
|
|
|
|
*
|
|
|
|
* @dev: Device to handle
|
|
|
|
* @next: impending interrupt/wakeup for the device
|
|
|
|
*
|
|
|
|
*
|
|
|
|
* Allow devices to inform of the next wakeup. It's assumed that the users
|
|
|
|
* guarantee that the genpd wouldn't be detached while this routine is getting
|
|
|
|
* called. Additionally, it's also assumed that @dev isn't runtime suspended
|
|
|
|
* (RPM_SUSPENDED)."
|
|
|
|
* Although devices are expected to update the next_wakeup after the end of
|
|
|
|
* their usecase as well, it is possible the devices themselves may not know
|
|
|
|
* about that, so stale @next will be ignored when powering off the domain.
|
|
|
|
*/
|
|
|
|
void dev_pm_genpd_set_next_wakeup(struct device *dev, ktime_t next)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
2022-05-11 14:56:57 +00:00
|
|
|
struct gpd_timing_data *td;
|
2021-01-20 15:50:41 +00:00
|
|
|
|
|
|
|
genpd = dev_to_genpd_safe(dev);
|
|
|
|
if (!genpd)
|
|
|
|
return;
|
|
|
|
|
2022-05-11 14:56:57 +00:00
|
|
|
td = to_gpd_data(dev->power.subsys_data->domain_data)->td;
|
|
|
|
if (td)
|
|
|
|
td->next_wakeup = next;
|
2021-01-20 15:50:41 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_set_next_wakeup);
|
|
|
|
|
2022-10-18 15:28:35 +00:00
|
|
|
/**
|
|
|
|
* dev_pm_genpd_get_next_hrtimer - Return the next_hrtimer for the genpd
|
|
|
|
* @dev: A device that is attached to the genpd.
|
|
|
|
*
|
|
|
|
* This routine should typically be called for a device, at the point of when a
|
|
|
|
* GENPD_NOTIFY_PRE_OFF notification has been sent for it.
|
|
|
|
*
|
|
|
|
* Returns the aggregated value of the genpd's next hrtimer or KTIME_MAX if no
|
|
|
|
* valid value have been set.
|
|
|
|
*/
|
|
|
|
ktime_t dev_pm_genpd_get_next_hrtimer(struct device *dev)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
|
|
|
|
|
|
|
genpd = dev_to_genpd_safe(dev);
|
|
|
|
if (!genpd)
|
|
|
|
return KTIME_MAX;
|
|
|
|
|
|
|
|
if (genpd->gd)
|
|
|
|
return genpd->gd->next_hrtimer;
|
|
|
|
|
|
|
|
return KTIME_MAX;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_get_next_hrtimer);
|
|
|
|
|
2023-01-02 10:48:27 +00:00
|
|
|
/*
|
|
|
|
* dev_pm_genpd_synced_poweroff - Next power off should be synchronous
|
|
|
|
*
|
|
|
|
* @dev: A device that is attached to the genpd.
|
|
|
|
*
|
|
|
|
* Allows a consumer of the genpd to notify the provider that the next power off
|
|
|
|
* should be synchronous.
|
|
|
|
*
|
|
|
|
* It is assumed that the users guarantee that the genpd wouldn't be detached
|
|
|
|
* while this routine is getting called.
|
|
|
|
*/
|
|
|
|
void dev_pm_genpd_synced_poweroff(struct device *dev)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
|
|
|
|
|
|
|
genpd = dev_to_genpd_safe(dev);
|
|
|
|
if (!genpd)
|
|
|
|
return;
|
|
|
|
|
|
|
|
genpd_lock(genpd);
|
|
|
|
genpd->synced_poweroff = true;
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_synced_poweroff);
|
|
|
|
|
PM: domains: Allow devices attached to genpd to be managed by HW
Some power-domains may be capable of relying on the HW to control the power
for a device that's hooked up to it. Typically, for these kinds of
configurations the consumer driver should be able to change the behavior of
power domain at runtime, control the power domain in SW mode for certain
configurations and handover the control to HW mode for other usecases.
To allow a consumer driver to change the behaviour of the PM domain for its
device, let's provide a new function, dev_pm_genpd_set_hwmode(). Moreover,
let's add a corresponding optional genpd callback, ->set_hwmode_dev(),
which the genpd provider should implement if it can support switching
between HW controlled mode and SW controlled mode. Similarly, add the
dev_pm_genpd_get_hwmode() to allow consumers to read the current mode and
its corresponding optional genpd callback, ->get_hwmode_dev(), which the
genpd provider can also implement to synchronize the initial HW mode
state in genpd_add_device() by reading back the mode from the hardware.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Abel Vesa <abel.vesa@linaro.org>
Signed-off-by: Jagadeesh Kona <quic_jkona@quicinc.com>
Reviewed-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
Reviewed-by: Dhruva Gole <d-gole@ti.com>
Reviewed-by: Taniya Das <quic_tdas@quicinc.com>
Link: https://lore.kernel.org/r/20240624044809.17751-2-quic_jkona@quicinc.com
2024-06-24 04:48:05 +00:00
|
|
|
/**
|
|
|
|
* dev_pm_genpd_set_hwmode() - Set the HW mode for the device and its PM domain.
|
|
|
|
*
|
|
|
|
* @dev: Device for which the HW-mode should be changed.
|
|
|
|
* @enable: Value to set or unset the HW-mode.
|
|
|
|
*
|
|
|
|
* Some PM domains can rely on HW signals to control the power for a device. To
|
|
|
|
* allow a consumer driver to switch the behaviour for its device in runtime,
|
|
|
|
* which may be beneficial from a latency or energy point of view, this function
|
|
|
|
* may be called.
|
|
|
|
*
|
|
|
|
* It is assumed that the users guarantee that the genpd wouldn't be detached
|
|
|
|
* while this routine is getting called.
|
|
|
|
*
|
|
|
|
* Return: Returns 0 on success and negative error values on failures.
|
|
|
|
*/
|
|
|
|
int dev_pm_genpd_set_hwmode(struct device *dev, bool enable)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
genpd = dev_to_genpd_safe(dev);
|
|
|
|
if (!genpd)
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
if (!genpd->set_hwmode_dev)
|
|
|
|
return -EOPNOTSUPP;
|
|
|
|
|
|
|
|
genpd_lock(genpd);
|
|
|
|
|
|
|
|
if (dev_gpd_data(dev)->hw_mode == enable)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
ret = genpd->set_hwmode_dev(genpd, dev, enable);
|
|
|
|
if (!ret)
|
|
|
|
dev_gpd_data(dev)->hw_mode = enable;
|
|
|
|
|
|
|
|
out:
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_set_hwmode);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* dev_pm_genpd_get_hwmode() - Get the HW mode setting for the device.
|
|
|
|
*
|
|
|
|
* @dev: Device for which the current HW-mode setting should be fetched.
|
|
|
|
*
|
|
|
|
* This helper function allows consumer drivers to fetch the current HW mode
|
|
|
|
* setting of its the device.
|
|
|
|
*
|
|
|
|
* It is assumed that the users guarantee that the genpd wouldn't be detached
|
|
|
|
* while this routine is getting called.
|
|
|
|
*
|
|
|
|
* Return: Returns the HW mode setting of device from SW cached hw_mode.
|
|
|
|
*/
|
|
|
|
bool dev_pm_genpd_get_hwmode(struct device *dev)
|
|
|
|
{
|
|
|
|
return dev_gpd_data(dev)->hw_mode;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_get_hwmode);
|
|
|
|
|
2016-12-08 13:45:20 +00:00
|
|
|
static int _genpd_power_on(struct generic_pm_domain *genpd, bool timed)
|
2014-11-10 18:39:19 +00:00
|
|
|
{
|
2016-02-15 10:10:51 +00:00
|
|
|
unsigned int state_idx = genpd->state_idx;
|
2014-11-10 18:39:19 +00:00
|
|
|
ktime_t time_start;
|
|
|
|
s64 elapsed_ns;
|
2020-10-20 08:10:35 +00:00
|
|
|
int ret;
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
|
|
|
|
/* Notify consumers that we are about to power on. */
|
2020-10-20 08:10:35 +00:00
|
|
|
ret = raw_notifier_call_chain_robust(&genpd->power_notifiers,
|
|
|
|
GENPD_NOTIFY_PRE_ON,
|
|
|
|
GENPD_NOTIFY_OFF, NULL);
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
ret = notifier_to_errno(ret);
|
|
|
|
if (ret)
|
2020-10-20 08:10:35 +00:00
|
|
|
return ret;
|
2014-11-10 18:39:19 +00:00
|
|
|
|
|
|
|
if (!genpd->power_on)
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
goto out;
|
2014-11-10 18:39:19 +00:00
|
|
|
|
2022-05-11 14:57:04 +00:00
|
|
|
timed = timed && genpd->gd && !genpd->states[state_idx].fwnode;
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
if (!timed) {
|
|
|
|
ret = genpd->power_on(genpd);
|
|
|
|
if (ret)
|
|
|
|
goto err;
|
|
|
|
|
|
|
|
goto out;
|
|
|
|
}
|
2015-05-29 15:24:23 +00:00
|
|
|
|
2014-11-10 18:39:19 +00:00
|
|
|
time_start = ktime_get();
|
|
|
|
ret = genpd->power_on(genpd);
|
|
|
|
if (ret)
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
goto err;
|
2014-11-10 18:39:19 +00:00
|
|
|
|
|
|
|
elapsed_ns = ktime_to_ns(ktime_sub(ktime_get(), time_start));
|
2016-02-15 10:10:51 +00:00
|
|
|
if (elapsed_ns <= genpd->states[state_idx].power_on_latency_ns)
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
goto out;
|
2014-11-10 18:39:19 +00:00
|
|
|
|
2016-02-15 10:10:51 +00:00
|
|
|
genpd->states[state_idx].power_on_latency_ns = elapsed_ns;
|
2022-05-11 14:57:03 +00:00
|
|
|
genpd->gd->max_off_time_changed = true;
|
2015-03-20 17:20:28 +00:00
|
|
|
pr_debug("%s: Power-%s latency exceeded, new value %lld ns\n",
|
2024-10-30 12:55:10 +00:00
|
|
|
dev_name(&genpd->dev), "on", elapsed_ns);
|
2014-11-10 18:39:19 +00:00
|
|
|
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
out:
|
|
|
|
raw_notifier_call_chain(&genpd->power_notifiers, GENPD_NOTIFY_ON, NULL);
|
2023-01-02 10:48:27 +00:00
|
|
|
genpd->synced_poweroff = false;
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
return 0;
|
|
|
|
err:
|
|
|
|
raw_notifier_call_chain(&genpd->power_notifiers, GENPD_NOTIFY_OFF,
|
|
|
|
NULL);
|
2014-11-10 18:39:19 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2016-12-08 13:45:20 +00:00
|
|
|
static int _genpd_power_off(struct generic_pm_domain *genpd, bool timed)
|
2014-11-10 18:39:19 +00:00
|
|
|
{
|
2016-02-15 10:10:51 +00:00
|
|
|
unsigned int state_idx = genpd->state_idx;
|
2014-11-10 18:39:19 +00:00
|
|
|
ktime_t time_start;
|
|
|
|
s64 elapsed_ns;
|
2020-10-20 08:10:35 +00:00
|
|
|
int ret;
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
|
|
|
|
/* Notify consumers that we are about to power off. */
|
2020-10-20 08:10:35 +00:00
|
|
|
ret = raw_notifier_call_chain_robust(&genpd->power_notifiers,
|
|
|
|
GENPD_NOTIFY_PRE_OFF,
|
|
|
|
GENPD_NOTIFY_ON, NULL);
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
ret = notifier_to_errno(ret);
|
|
|
|
if (ret)
|
2020-10-20 08:10:35 +00:00
|
|
|
return ret;
|
2014-11-10 18:39:19 +00:00
|
|
|
|
|
|
|
if (!genpd->power_off)
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
goto out;
|
|
|
|
|
2022-05-11 14:57:04 +00:00
|
|
|
timed = timed && genpd->gd && !genpd->states[state_idx].fwnode;
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
if (!timed) {
|
|
|
|
ret = genpd->power_off(genpd);
|
|
|
|
if (ret)
|
|
|
|
goto busy;
|
2014-11-10 18:39:19 +00:00
|
|
|
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
goto out;
|
|
|
|
}
|
2015-05-29 15:24:23 +00:00
|
|
|
|
2014-11-10 18:39:19 +00:00
|
|
|
time_start = ktime_get();
|
|
|
|
ret = genpd->power_off(genpd);
|
2019-03-06 13:25:15 +00:00
|
|
|
if (ret)
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
goto busy;
|
2014-11-10 18:39:19 +00:00
|
|
|
|
|
|
|
elapsed_ns = ktime_to_ns(ktime_sub(ktime_get(), time_start));
|
2016-02-15 10:10:51 +00:00
|
|
|
if (elapsed_ns <= genpd->states[state_idx].power_off_latency_ns)
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
goto out;
|
2014-11-10 18:39:19 +00:00
|
|
|
|
2016-02-15 10:10:51 +00:00
|
|
|
genpd->states[state_idx].power_off_latency_ns = elapsed_ns;
|
2022-05-11 14:57:03 +00:00
|
|
|
genpd->gd->max_off_time_changed = true;
|
2015-03-20 17:20:28 +00:00
|
|
|
pr_debug("%s: Power-%s latency exceeded, new value %lld ns\n",
|
2024-10-30 12:55:10 +00:00
|
|
|
dev_name(&genpd->dev), "off", elapsed_ns);
|
2014-11-10 18:39:19 +00:00
|
|
|
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
out:
|
|
|
|
raw_notifier_call_chain(&genpd->power_notifiers, GENPD_NOTIFY_OFF,
|
|
|
|
NULL);
|
2019-03-06 13:25:15 +00:00
|
|
|
return 0;
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
busy:
|
2020-10-20 08:10:35 +00:00
|
|
|
raw_notifier_call_chain(&genpd->power_notifiers, GENPD_NOTIFY_ON, NULL);
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
return ret;
|
2014-11-10 18:39:19 +00:00
|
|
|
}
|
|
|
|
|
2015-09-02 08:16:13 +00:00
|
|
|
/**
|
2016-12-08 13:45:20 +00:00
|
|
|
* genpd_queue_power_off_work - Queue up the execution of genpd_power_off().
|
2016-01-27 07:29:27 +00:00
|
|
|
* @genpd: PM domain to power off.
|
2015-09-02 08:16:13 +00:00
|
|
|
*
|
2016-12-08 13:45:20 +00:00
|
|
|
* Queue up the execution of genpd_power_off() unless it's already been done
|
2015-09-02 08:16:13 +00:00
|
|
|
* before.
|
|
|
|
*/
|
|
|
|
static void genpd_queue_power_off_work(struct generic_pm_domain *genpd)
|
|
|
|
{
|
|
|
|
queue_work(pm_wq, &genpd->power_off_work);
|
|
|
|
}
|
|
|
|
|
2017-02-17 09:55:23 +00:00
|
|
|
/**
|
|
|
|
* genpd_power_off - Remove power from a given PM domain.
|
|
|
|
* @genpd: PM domain to power down.
|
2017-02-17 09:55:24 +00:00
|
|
|
* @one_dev_on: If invoked from genpd's ->runtime_suspend|resume() callback, the
|
|
|
|
* RPM status of the releated device is in an intermediate state, not yet turned
|
|
|
|
* into RPM_SUSPENDED. This means genpd_power_off() must allow one device to not
|
|
|
|
* be RPM_SUSPENDED, while it tries to power off the PM domain.
|
2021-05-12 07:25:15 +00:00
|
|
|
* @depth: nesting count for lockdep.
|
2017-02-17 09:55:23 +00:00
|
|
|
*
|
|
|
|
* If all of the @genpd's devices have been suspended and all of its subdomains
|
|
|
|
* have been powered down, remove power from @genpd.
|
|
|
|
*/
|
2017-02-17 09:55:25 +00:00
|
|
|
static int genpd_power_off(struct generic_pm_domain *genpd, bool one_dev_on,
|
|
|
|
unsigned int depth)
|
2017-02-17 09:55:23 +00:00
|
|
|
{
|
|
|
|
struct pm_domain_data *pdd;
|
|
|
|
struct gpd_link *link;
|
|
|
|
unsigned int not_suspended = 0;
|
2020-09-24 11:04:48 +00:00
|
|
|
int ret;
|
2017-02-17 09:55:23 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Do not try to power off the domain in the following situations:
|
|
|
|
* (1) The domain is already in the "power off" state.
|
|
|
|
* (2) System suspend is in progress.
|
|
|
|
*/
|
2017-03-20 10:19:20 +00:00
|
|
|
if (!genpd_status_on(genpd) || genpd->prepared_count > 0)
|
2017-02-17 09:55:23 +00:00
|
|
|
return 0;
|
|
|
|
|
2017-03-20 10:19:21 +00:00
|
|
|
/*
|
|
|
|
* Abort power off for the PM domain in the following situations:
|
|
|
|
* (1) The domain is configured as always on.
|
|
|
|
* (2) When the domain has a subdomain being powered on.
|
|
|
|
*/
|
2019-04-30 15:06:11 +00:00
|
|
|
if (genpd_is_always_on(genpd) ||
|
|
|
|
genpd_is_rpm_always_on(genpd) ||
|
|
|
|
atomic_read(&genpd->sd_count) > 0)
|
2017-02-17 09:55:23 +00:00
|
|
|
return -EBUSY;
|
|
|
|
|
2022-02-17 12:49:50 +00:00
|
|
|
/*
|
|
|
|
* The children must be in their deepest (powered-off) states to allow
|
|
|
|
* the parent to be powered off. Note that, there's no need for
|
|
|
|
* additional locking, as powering on a child, requires the parent's
|
|
|
|
* lock to be acquired first.
|
|
|
|
*/
|
|
|
|
list_for_each_entry(link, &genpd->parent_links, parent_node) {
|
|
|
|
struct generic_pm_domain *child = link->child;
|
|
|
|
if (child->state_idx < child->state_count - 1)
|
|
|
|
return -EBUSY;
|
|
|
|
}
|
|
|
|
|
2017-02-17 09:55:23 +00:00
|
|
|
list_for_each_entry(pdd, &genpd->dev_list, list_node) {
|
|
|
|
/*
|
|
|
|
* Do not allow PM domain to be powered off, when an IRQ safe
|
|
|
|
* device is part of a non-IRQ safe domain.
|
|
|
|
*/
|
|
|
|
if (!pm_runtime_suspended(pdd->dev) ||
|
2022-05-11 14:56:54 +00:00
|
|
|
irq_safe_dev_in_sleep_domain(pdd->dev, genpd))
|
2017-02-17 09:55:23 +00:00
|
|
|
not_suspended++;
|
|
|
|
}
|
|
|
|
|
2017-02-17 09:55:24 +00:00
|
|
|
if (not_suspended > 1 || (not_suspended == 1 && !one_dev_on))
|
2017-02-17 09:55:23 +00:00
|
|
|
return -EBUSY;
|
|
|
|
|
|
|
|
if (genpd->gov && genpd->gov->power_down_ok) {
|
|
|
|
if (!genpd->gov->power_down_ok(&genpd->domain))
|
|
|
|
return -EAGAIN;
|
|
|
|
}
|
|
|
|
|
2018-10-03 14:38:15 +00:00
|
|
|
/* Default to shallowest state. */
|
|
|
|
if (!genpd->gov)
|
|
|
|
genpd->state_idx = 0;
|
|
|
|
|
2020-09-24 11:04:48 +00:00
|
|
|
/* Don't power off, if a child domain is waiting to power on. */
|
|
|
|
if (atomic_read(&genpd->sd_count) > 0)
|
|
|
|
return -EBUSY;
|
2017-02-17 09:55:23 +00:00
|
|
|
|
2020-09-24 11:04:48 +00:00
|
|
|
ret = _genpd_power_off(genpd, true);
|
2020-10-15 20:47:22 +00:00
|
|
|
if (ret) {
|
|
|
|
genpd->states[genpd->state_idx].rejected++;
|
2020-09-24 11:04:48 +00:00
|
|
|
return ret;
|
2020-10-15 20:47:22 +00:00
|
|
|
}
|
2017-02-17 09:55:23 +00:00
|
|
|
|
2020-09-24 11:04:47 +00:00
|
|
|
genpd->status = GENPD_STATE_OFF;
|
2017-07-14 17:10:15 +00:00
|
|
|
genpd_update_accounting(genpd);
|
2020-10-15 20:47:22 +00:00
|
|
|
genpd->states[genpd->state_idx].usage++;
|
2017-02-17 09:55:23 +00:00
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry(link, &genpd->child_links, child_node) {
|
|
|
|
genpd_sd_counter_dec(link->parent);
|
|
|
|
genpd_lock_nested(link->parent, depth + 1);
|
|
|
|
genpd_power_off(link->parent, false, depth + 1);
|
|
|
|
genpd_unlock(link->parent);
|
2017-02-17 09:55:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2011-07-01 20:13:10 +00:00
|
|
|
/**
|
2020-07-08 23:32:13 +00:00
|
|
|
* genpd_power_on - Restore power to a given PM domain and its parents.
|
2011-07-01 20:13:10 +00:00
|
|
|
* @genpd: PM domain to power up.
|
2016-01-20 09:13:42 +00:00
|
|
|
* @depth: nesting count for lockdep.
|
2011-07-01 20:13:10 +00:00
|
|
|
*
|
2020-07-08 23:32:13 +00:00
|
|
|
* Restore power to @genpd and all of its parents so that it is possible to
|
2011-07-01 20:13:10 +00:00
|
|
|
* resume a device belonging to it.
|
|
|
|
*/
|
2016-12-08 13:45:20 +00:00
|
|
|
static int genpd_power_on(struct generic_pm_domain *genpd, unsigned int depth)
|
2011-07-01 20:13:10 +00:00
|
|
|
{
|
2011-08-08 21:43:40 +00:00
|
|
|
struct gpd_link *link;
|
2011-07-01 20:13:10 +00:00
|
|
|
int ret = 0;
|
|
|
|
|
2017-03-20 10:19:20 +00:00
|
|
|
if (genpd_status_on(genpd))
|
2011-08-08 21:43:29 +00:00
|
|
|
return 0;
|
2011-07-01 20:13:10 +00:00
|
|
|
|
2011-08-08 21:43:40 +00:00
|
|
|
/*
|
|
|
|
* The list is guaranteed not to change while the loop below is being
|
2020-07-08 23:32:13 +00:00
|
|
|
* executed, unless one of the parents' .power_on() callbacks fiddles
|
2011-08-08 21:43:40 +00:00
|
|
|
* with it.
|
|
|
|
*/
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry(link, &genpd->child_links, child_node) {
|
|
|
|
struct generic_pm_domain *parent = link->parent;
|
2016-01-20 09:13:42 +00:00
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_sd_counter_inc(parent);
|
2016-01-20 09:13:42 +00:00
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_lock_nested(parent, depth + 1);
|
|
|
|
ret = genpd_power_on(parent, depth + 1);
|
|
|
|
genpd_unlock(parent);
|
2011-07-01 20:13:10 +00:00
|
|
|
|
2011-08-08 21:43:40 +00:00
|
|
|
if (ret) {
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_sd_counter_dec(parent);
|
2011-08-08 21:43:22 +00:00
|
|
|
goto err;
|
2011-08-08 21:43:40 +00:00
|
|
|
}
|
2011-07-01 20:13:10 +00:00
|
|
|
}
|
|
|
|
|
2016-12-08 13:45:20 +00:00
|
|
|
ret = _genpd_power_on(genpd, true);
|
2014-11-10 18:39:19 +00:00
|
|
|
if (ret)
|
|
|
|
goto err;
|
2011-07-01 20:13:10 +00:00
|
|
|
|
2020-09-24 11:04:47 +00:00
|
|
|
genpd->status = GENPD_STATE_ON;
|
2017-07-14 17:10:15 +00:00
|
|
|
genpd_update_accounting(genpd);
|
|
|
|
|
2011-08-08 21:43:29 +00:00
|
|
|
return 0;
|
2011-08-08 21:43:22 +00:00
|
|
|
|
|
|
|
err:
|
2015-09-02 08:16:13 +00:00
|
|
|
list_for_each_entry_continue_reverse(link,
|
2020-07-08 23:32:13 +00:00
|
|
|
&genpd->child_links,
|
|
|
|
child_node) {
|
|
|
|
genpd_sd_counter_dec(link->parent);
|
|
|
|
genpd_lock_nested(link->parent, depth + 1);
|
|
|
|
genpd_power_off(link->parent, false, depth + 1);
|
|
|
|
genpd_unlock(link->parent);
|
2015-09-02 08:16:13 +00:00
|
|
|
}
|
2011-08-08 21:43:22 +00:00
|
|
|
|
2011-08-08 21:43:29 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2019-10-16 13:16:24 +00:00
|
|
|
static int genpd_dev_pm_start(struct device *dev)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = dev_to_genpd(dev);
|
|
|
|
|
|
|
|
return genpd_start_dev(genpd, dev);
|
|
|
|
}
|
|
|
|
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
static int genpd_dev_pm_qos_notifier(struct notifier_block *nb,
|
|
|
|
unsigned long val, void *ptr)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain_data *gpd_data;
|
|
|
|
struct device *dev;
|
|
|
|
|
|
|
|
gpd_data = container_of(nb, struct generic_pm_domain_data, nb);
|
|
|
|
dev = gpd_data->base.dev;
|
|
|
|
|
|
|
|
for (;;) {
|
2022-05-11 14:57:02 +00:00
|
|
|
struct generic_pm_domain *genpd = ERR_PTR(-ENODATA);
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
struct pm_domain_data *pdd;
|
2022-05-11 14:56:56 +00:00
|
|
|
struct gpd_timing_data *td;
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
|
|
|
|
spin_lock_irq(&dev->power.lock);
|
|
|
|
|
|
|
|
pdd = dev->power.subsys_data ?
|
|
|
|
dev->power.subsys_data->domain_data : NULL;
|
2017-05-16 05:22:43 +00:00
|
|
|
if (pdd) {
|
2022-05-11 14:56:56 +00:00
|
|
|
td = to_gpd_data(pdd)->td;
|
2022-05-11 14:57:02 +00:00
|
|
|
if (td) {
|
2022-05-11 14:56:56 +00:00
|
|
|
td->constraint_changed = true;
|
2022-05-11 14:57:02 +00:00
|
|
|
genpd = dev_to_genpd(dev);
|
|
|
|
}
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
spin_unlock_irq(&dev->power.lock);
|
|
|
|
|
|
|
|
if (!IS_ERR(genpd)) {
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
2022-05-11 14:57:02 +00:00
|
|
|
genpd->gd->max_off_time_changed = true;
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
dev = dev->parent;
|
|
|
|
if (!dev || dev->power.ignore_children)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return NOTIFY_DONE;
|
|
|
|
}
|
|
|
|
|
2011-07-01 20:12:45 +00:00
|
|
|
/**
|
|
|
|
* genpd_power_off_work_fn - Power off PM domain whose subdomain count is 0.
|
|
|
|
* @work: Work structure used for scheduling the execution of this function.
|
|
|
|
*/
|
|
|
|
static void genpd_power_off_work_fn(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
|
|
|
|
|
|
|
genpd = container_of(work, struct generic_pm_domain, power_off_work);
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
2017-02-17 09:55:25 +00:00
|
|
|
genpd_power_off(genpd, false, 0);
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2011-07-01 20:12:45 +00:00
|
|
|
}
|
|
|
|
|
2016-03-31 09:21:27 +00:00
|
|
|
/**
|
|
|
|
* __genpd_runtime_suspend - walk the hierarchy of ->runtime_suspend() callbacks
|
|
|
|
* @dev: Device to handle.
|
|
|
|
*/
|
|
|
|
static int __genpd_runtime_suspend(struct device *dev)
|
|
|
|
{
|
|
|
|
int (*cb)(struct device *__dev);
|
|
|
|
|
|
|
|
if (dev->type && dev->type->pm)
|
|
|
|
cb = dev->type->pm->runtime_suspend;
|
|
|
|
else if (dev->class && dev->class->pm)
|
|
|
|
cb = dev->class->pm->runtime_suspend;
|
|
|
|
else if (dev->bus && dev->bus->pm)
|
|
|
|
cb = dev->bus->pm->runtime_suspend;
|
|
|
|
else
|
|
|
|
cb = NULL;
|
|
|
|
|
|
|
|
if (!cb && dev->driver && dev->driver->pm)
|
|
|
|
cb = dev->driver->pm->runtime_suspend;
|
|
|
|
|
|
|
|
return cb ? cb(dev) : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* __genpd_runtime_resume - walk the hierarchy of ->runtime_resume() callbacks
|
|
|
|
* @dev: Device to handle.
|
|
|
|
*/
|
|
|
|
static int __genpd_runtime_resume(struct device *dev)
|
|
|
|
{
|
|
|
|
int (*cb)(struct device *__dev);
|
|
|
|
|
|
|
|
if (dev->type && dev->type->pm)
|
|
|
|
cb = dev->type->pm->runtime_resume;
|
|
|
|
else if (dev->class && dev->class->pm)
|
|
|
|
cb = dev->class->pm->runtime_resume;
|
|
|
|
else if (dev->bus && dev->bus->pm)
|
|
|
|
cb = dev->bus->pm->runtime_resume;
|
|
|
|
else
|
|
|
|
cb = NULL;
|
|
|
|
|
|
|
|
if (!cb && dev->driver && dev->driver->pm)
|
|
|
|
cb = dev->driver->pm->runtime_resume;
|
|
|
|
|
|
|
|
return cb ? cb(dev) : 0;
|
|
|
|
}
|
|
|
|
|
2011-07-01 20:12:45 +00:00
|
|
|
/**
|
2016-03-31 09:21:26 +00:00
|
|
|
* genpd_runtime_suspend - Suspend a device belonging to I/O PM domain.
|
2011-07-01 20:12:45 +00:00
|
|
|
* @dev: Device to suspend.
|
|
|
|
*
|
|
|
|
* Carry out a runtime suspend of a device under the assumption that its
|
|
|
|
* pm_domain field points to the domain member of an object of type
|
|
|
|
* struct generic_pm_domain representing a PM domain consisting of I/O devices.
|
|
|
|
*/
|
2016-03-31 09:21:26 +00:00
|
|
|
static int genpd_runtime_suspend(struct device *dev)
|
2011-07-01 20:12:45 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
2016-03-31 09:21:25 +00:00
|
|
|
bool (*suspend_ok)(struct device *__dev);
|
2021-06-03 09:34:37 +00:00
|
|
|
struct generic_pm_domain_data *gpd_data = dev_gpd_data(dev);
|
2022-05-11 14:56:56 +00:00
|
|
|
struct gpd_timing_data *td = gpd_data->td;
|
2015-11-30 15:21:38 +00:00
|
|
|
bool runtime_pm = pm_runtime_enabled(dev);
|
2022-05-11 14:56:58 +00:00
|
|
|
ktime_t time_start = 0;
|
2015-10-15 15:02:19 +00:00
|
|
|
s64 elapsed_ns;
|
2011-11-27 12:11:36 +00:00
|
|
|
int ret;
|
2011-07-01 20:12:45 +00:00
|
|
|
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
2011-07-01 20:13:10 +00:00
|
|
|
genpd = dev_to_genpd(dev);
|
|
|
|
if (IS_ERR(genpd))
|
2011-07-01 20:12:45 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
2015-11-30 15:21:38 +00:00
|
|
|
/*
|
|
|
|
* A runtime PM centric subsystem/driver may re-use the runtime PM
|
|
|
|
* callbacks for other purposes than runtime PM. In those scenarios
|
|
|
|
* runtime PM is disabled. Under these circumstances, we shall skip
|
|
|
|
* validating/measuring the PM QoS latency.
|
|
|
|
*/
|
2016-03-31 09:21:25 +00:00
|
|
|
suspend_ok = genpd->gov ? genpd->gov->suspend_ok : NULL;
|
|
|
|
if (runtime_pm && suspend_ok && !suspend_ok(dev))
|
2011-11-30 23:02:05 +00:00
|
|
|
return -EBUSY;
|
|
|
|
|
2015-10-15 15:02:19 +00:00
|
|
|
/* Measure suspend latency. */
|
2022-05-11 14:56:58 +00:00
|
|
|
if (td && runtime_pm)
|
2015-11-30 15:21:38 +00:00
|
|
|
time_start = ktime_get();
|
2015-10-15 15:02:19 +00:00
|
|
|
|
2016-03-31 09:21:27 +00:00
|
|
|
ret = __genpd_runtime_suspend(dev);
|
2011-11-27 12:11:36 +00:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
2011-07-11 22:39:29 +00:00
|
|
|
|
2015-10-15 15:02:19 +00:00
|
|
|
ret = genpd_stop_dev(genpd, dev);
|
PM / Domains: Remove intermediate states from the power off sequence
Genpd's ->runtime_suspend() (assigned to pm_genpd_runtime_suspend())
doesn't immediately walk the hierarchy of ->runtime_suspend() callbacks.
Instead, pm_genpd_runtime_suspend() calls pm_genpd_poweroff() which
postpones that until *all* the devices in the genpd are runtime suspended.
When pm_genpd_poweroff() discovers that the last device in the genpd is
about to be runtime suspended, it calls __pm_genpd_save_device() for *all*
the devices in the genpd sequentially. Furthermore,
__pm_genpd_save_device() invokes the ->start() callback, walks the
hierarchy of the ->runtime_suspend() callbacks and invokes the ->stop()
callback. This causes a "thundering herd" problem.
Let's address this issue by having pm_genpd_runtime_suspend() immediately
walk the hierarchy of the ->runtime_suspend() callbacks, instead of
postponing that to the power off sequence via pm_genpd_poweroff(). If the
selected ->runtime_suspend() callback doesn't return an error code, call
pm_genpd_poweroff() to see if it's feasible to also power off the PM
domain.
Adopting this change enables us to simplify parts of the code in genpd,
for example the locking mechanism. Additionally, it gives some positive
side effects, as described below.
i)
One device's ->runtime_resume() latency is no longer affected by other
devices' latencies in a genpd.
The complexity genpd has to support the option to abort the power off
sequence suffers from latency issues. More precisely, a device that is
requested to be runtime resumed, may end up waiting for
__pm_genpd_save_device() to complete its operations for *another* device.
That's because pm_genpd_poweroff() can't confirm an abort request while it
waits for __pm_genpd_save_device() to return.
As this patch removes the intermediate states in pm_genpd_poweroff() while
powering off the PM domain, we no longer need the ability to abort that
sequence.
ii)
Make pm_runtime[_status]_suspended() reliable when used with genpd.
Until the last device in a genpd becomes idle, pm_genpd_runtime_suspend()
will return 0 without actually walking the hierarchy of the
->runtime_suspend() callbacks. However, by returning 0 the runtime PM core
considers the device as runtime_suspended, so
pm_runtime[_status]_suspended() will return true, even though the device
isn't (yet) runtime suspended.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks,
pm_runtime[_status]_suspended() will accurately reflect the status of the
device.
iii)
Enable fine-grained PM through runtime PM callbacks in drivers/subsystems.
There are currently cases were drivers/subsystems implements runtime PM
callbacks to deploy fine-grained PM (e.g. gate clocks, move pinctrl to
power-save state, etc.). While using the genpd, pm_genpd_runtime_suspend()
postpones invoking these callbacks until *all* the devices in the genpd
are runtime suspended. In essence, one runtime resumed device prevents
fine-grained PM for other devices within the same genpd.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks, fine-grained PM is enabled
throughout all the levels of runtime PM callbacks.
iiii)
Enable fine-grained PM for IRQ safe devices
Per the definition for an IRQ safe device, its runtime PM callbacks must
be able to execute in atomic context. In the path while genpd walks the
hierarchy of the ->runtime_suspend() callbacks for the device, it uses a
mutex. Therefore, genpd prevents that path to be executed for IRQ safe
devices.
As this patch changes pm_genpd_runtime_suspend() to immediately walk the
hierarchy of the ->runtime_suspend() callbacks and without needing to use
a mutex, fine-grained PM is enabled throughout all the levels of runtime
PM callbacks for IRQ safe devices.
Unfortunately this patch also comes with a drawback, as described in the
summary below.
Driver's/subsystem's runtime PM callbacks may be invoked even when the
genpd hasn't actually powered off the PM domain, potentially introducing
unnecessary latency.
However, in most cases, saving/restoring register contexts for devices are
typically fast operations or can be optimized in device specific ways
(e.g. shadow copies of register contents in memory, device-specific checks
to see if context has been lost before restoring context, etc.).
Still, in some cases the driver/subsystem may suffer from latency if
runtime PM is used in a very fine-grained manner (e.g. for each IO request
or xfer). To prevent that extra overhead, the driver/subsystem may deploy
the runtime PM autosuspend feature.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Reviewed-by: Kevin Hilman <khilman@linaro.org>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Tested-by: Lina Iyer <lina.iyer@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2015-06-18 13:17:53 +00:00
|
|
|
if (ret) {
|
2016-03-31 09:21:27 +00:00
|
|
|
__genpd_runtime_resume(dev);
|
PM / Domains: Remove intermediate states from the power off sequence
Genpd's ->runtime_suspend() (assigned to pm_genpd_runtime_suspend())
doesn't immediately walk the hierarchy of ->runtime_suspend() callbacks.
Instead, pm_genpd_runtime_suspend() calls pm_genpd_poweroff() which
postpones that until *all* the devices in the genpd are runtime suspended.
When pm_genpd_poweroff() discovers that the last device in the genpd is
about to be runtime suspended, it calls __pm_genpd_save_device() for *all*
the devices in the genpd sequentially. Furthermore,
__pm_genpd_save_device() invokes the ->start() callback, walks the
hierarchy of the ->runtime_suspend() callbacks and invokes the ->stop()
callback. This causes a "thundering herd" problem.
Let's address this issue by having pm_genpd_runtime_suspend() immediately
walk the hierarchy of the ->runtime_suspend() callbacks, instead of
postponing that to the power off sequence via pm_genpd_poweroff(). If the
selected ->runtime_suspend() callback doesn't return an error code, call
pm_genpd_poweroff() to see if it's feasible to also power off the PM
domain.
Adopting this change enables us to simplify parts of the code in genpd,
for example the locking mechanism. Additionally, it gives some positive
side effects, as described below.
i)
One device's ->runtime_resume() latency is no longer affected by other
devices' latencies in a genpd.
The complexity genpd has to support the option to abort the power off
sequence suffers from latency issues. More precisely, a device that is
requested to be runtime resumed, may end up waiting for
__pm_genpd_save_device() to complete its operations for *another* device.
That's because pm_genpd_poweroff() can't confirm an abort request while it
waits for __pm_genpd_save_device() to return.
As this patch removes the intermediate states in pm_genpd_poweroff() while
powering off the PM domain, we no longer need the ability to abort that
sequence.
ii)
Make pm_runtime[_status]_suspended() reliable when used with genpd.
Until the last device in a genpd becomes idle, pm_genpd_runtime_suspend()
will return 0 without actually walking the hierarchy of the
->runtime_suspend() callbacks. However, by returning 0 the runtime PM core
considers the device as runtime_suspended, so
pm_runtime[_status]_suspended() will return true, even though the device
isn't (yet) runtime suspended.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks,
pm_runtime[_status]_suspended() will accurately reflect the status of the
device.
iii)
Enable fine-grained PM through runtime PM callbacks in drivers/subsystems.
There are currently cases were drivers/subsystems implements runtime PM
callbacks to deploy fine-grained PM (e.g. gate clocks, move pinctrl to
power-save state, etc.). While using the genpd, pm_genpd_runtime_suspend()
postpones invoking these callbacks until *all* the devices in the genpd
are runtime suspended. In essence, one runtime resumed device prevents
fine-grained PM for other devices within the same genpd.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks, fine-grained PM is enabled
throughout all the levels of runtime PM callbacks.
iiii)
Enable fine-grained PM for IRQ safe devices
Per the definition for an IRQ safe device, its runtime PM callbacks must
be able to execute in atomic context. In the path while genpd walks the
hierarchy of the ->runtime_suspend() callbacks for the device, it uses a
mutex. Therefore, genpd prevents that path to be executed for IRQ safe
devices.
As this patch changes pm_genpd_runtime_suspend() to immediately walk the
hierarchy of the ->runtime_suspend() callbacks and without needing to use
a mutex, fine-grained PM is enabled throughout all the levels of runtime
PM callbacks for IRQ safe devices.
Unfortunately this patch also comes with a drawback, as described in the
summary below.
Driver's/subsystem's runtime PM callbacks may be invoked even when the
genpd hasn't actually powered off the PM domain, potentially introducing
unnecessary latency.
However, in most cases, saving/restoring register contexts for devices are
typically fast operations or can be optimized in device specific ways
(e.g. shadow copies of register contents in memory, device-specific checks
to see if context has been lost before restoring context, etc.).
Still, in some cases the driver/subsystem may suffer from latency if
runtime PM is used in a very fine-grained manner (e.g. for each IO request
or xfer). To prevent that extra overhead, the driver/subsystem may deploy
the runtime PM autosuspend feature.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Reviewed-by: Kevin Hilman <khilman@linaro.org>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Tested-by: Lina Iyer <lina.iyer@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2015-06-18 13:17:53 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2015-10-15 15:02:19 +00:00
|
|
|
/* Update suspend latency value if the measured time exceeds it. */
|
2022-05-11 14:56:58 +00:00
|
|
|
if (td && runtime_pm) {
|
2015-11-30 15:21:38 +00:00
|
|
|
elapsed_ns = ktime_to_ns(ktime_sub(ktime_get(), time_start));
|
2022-05-11 14:56:58 +00:00
|
|
|
if (elapsed_ns > td->suspend_latency_ns) {
|
2015-11-30 15:21:38 +00:00
|
|
|
td->suspend_latency_ns = elapsed_ns;
|
|
|
|
dev_dbg(dev, "suspend latency exceeded, %lld ns\n",
|
|
|
|
elapsed_ns);
|
2022-05-11 14:57:02 +00:00
|
|
|
genpd->gd->max_off_time_changed = true;
|
2015-11-30 15:21:38 +00:00
|
|
|
td->constraint_changed = true;
|
|
|
|
}
|
2015-10-15 15:02:19 +00:00
|
|
|
}
|
|
|
|
|
2011-08-25 13:37:04 +00:00
|
|
|
/*
|
2016-10-14 17:47:55 +00:00
|
|
|
* If power.irq_safe is set, this routine may be run with
|
|
|
|
* IRQs disabled, so suspend only if the PM domain also is irq_safe.
|
2011-08-25 13:37:04 +00:00
|
|
|
*/
|
2022-05-11 14:56:54 +00:00
|
|
|
if (irq_safe_dev_in_sleep_domain(dev, genpd))
|
2011-08-25 13:37:04 +00:00
|
|
|
return 0;
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
2017-02-17 09:55:25 +00:00
|
|
|
genpd_power_off(genpd, true, 0);
|
PM: domains: Reverse the order of performance and enabling ops
The ->set_performance_state() needs to be called before ->power_on()
when a genpd is powered on, and after ->power_off() when a genpd is
powered off. Do this in order to let the provider know to which
performance state to power on the genpd, on the power on sequence, and
also to maintain the performance for that genpd until after powering off,
on power off sequence.
There is no scenario where a consumer would need its genpd enabled and
then its performance state increased. Instead, in every scenario, the
consumer needs the genpd to be enabled from the start at a specific
performance state.
And same logic applies to the powering down. No consumer would need its
genpd performance state dropped right before powering down.
Now, there are currently two vendors which use ->set_performance_state()
in their genpd providers. One of them is Tegra, but the only genpd provider
(PMC) that makes use of ->set_performance_state() doesn't implement the
->power_on() or ->power_off(), and so it will not be affected by the ops
reversal.
The other vendor that uses it is Qualcomm, in multiple genpd providers
actually (RPM, RPMh and CPR). But all Qualcomm genpd providers that make
use of ->set_performance_state() need the order between enabling ops and
the performance setting op to be reversed. And the reason for that is that
it currently translates into two different voltages in order to power on
a genpd to a specific performance state. Basically, ->power_on() switches
to the minimum (enabling) voltage for that genpd, and then
->set_performance_state() sets it to the voltage level required by the
consumer.
By reversing the call order, we rely on the provider to know what to do
on each call, but most popular usecase is to cache the performance state
and postpone the voltage setting until the ->power_on() gets called.
As for the reason of still needing the ->power_on() and ->power_off() for a
provider which could get away with just having ->set_performance_state()
implemented, there are consumers that do not (nor should) provide an
opp-table. For those consumers, ->set_performance_state() will not be
called, and so they will enable the genpd to its minimum performance state
by a ->power_on() call. Same logic goes for the disabling.
Signed-off-by: Abel Vesa <abel.vesa@linaro.org>
Reviewed-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2022-11-15 21:25:43 +00:00
|
|
|
gpd_data->rpm_pstate = genpd_drop_performance_state(dev);
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2016-03-31 09:21:26 +00:00
|
|
|
* genpd_runtime_resume - Resume a device belonging to I/O PM domain.
|
2011-07-01 20:12:45 +00:00
|
|
|
* @dev: Device to resume.
|
|
|
|
*
|
|
|
|
* Carry out a runtime resume of a device under the assumption that its
|
|
|
|
* pm_domain field points to the domain member of an object of type
|
|
|
|
* struct generic_pm_domain representing a PM domain consisting of I/O devices.
|
|
|
|
*/
|
2016-03-31 09:21:26 +00:00
|
|
|
static int genpd_runtime_resume(struct device *dev)
|
2011-07-01 20:12:45 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
2021-06-03 09:34:37 +00:00
|
|
|
struct generic_pm_domain_data *gpd_data = dev_gpd_data(dev);
|
2022-05-11 14:56:56 +00:00
|
|
|
struct gpd_timing_data *td = gpd_data->td;
|
2022-05-11 14:56:58 +00:00
|
|
|
bool timed = td && pm_runtime_enabled(dev);
|
|
|
|
ktime_t time_start = 0;
|
2015-10-15 15:02:19 +00:00
|
|
|
s64 elapsed_ns;
|
2011-07-01 20:12:45 +00:00
|
|
|
int ret;
|
|
|
|
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
2011-07-01 20:13:10 +00:00
|
|
|
genpd = dev_to_genpd(dev);
|
|
|
|
if (IS_ERR(genpd))
|
2011-07-01 20:12:45 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
2016-10-14 17:47:55 +00:00
|
|
|
/*
|
|
|
|
* As we don't power off a non IRQ safe domain, which holds
|
|
|
|
* an IRQ safe device, we don't need to restore power to it.
|
|
|
|
*/
|
2022-05-11 14:56:59 +00:00
|
|
|
if (irq_safe_dev_in_sleep_domain(dev, genpd))
|
PM / Domains: Remove intermediate states from the power off sequence
Genpd's ->runtime_suspend() (assigned to pm_genpd_runtime_suspend())
doesn't immediately walk the hierarchy of ->runtime_suspend() callbacks.
Instead, pm_genpd_runtime_suspend() calls pm_genpd_poweroff() which
postpones that until *all* the devices in the genpd are runtime suspended.
When pm_genpd_poweroff() discovers that the last device in the genpd is
about to be runtime suspended, it calls __pm_genpd_save_device() for *all*
the devices in the genpd sequentially. Furthermore,
__pm_genpd_save_device() invokes the ->start() callback, walks the
hierarchy of the ->runtime_suspend() callbacks and invokes the ->stop()
callback. This causes a "thundering herd" problem.
Let's address this issue by having pm_genpd_runtime_suspend() immediately
walk the hierarchy of the ->runtime_suspend() callbacks, instead of
postponing that to the power off sequence via pm_genpd_poweroff(). If the
selected ->runtime_suspend() callback doesn't return an error code, call
pm_genpd_poweroff() to see if it's feasible to also power off the PM
domain.
Adopting this change enables us to simplify parts of the code in genpd,
for example the locking mechanism. Additionally, it gives some positive
side effects, as described below.
i)
One device's ->runtime_resume() latency is no longer affected by other
devices' latencies in a genpd.
The complexity genpd has to support the option to abort the power off
sequence suffers from latency issues. More precisely, a device that is
requested to be runtime resumed, may end up waiting for
__pm_genpd_save_device() to complete its operations for *another* device.
That's because pm_genpd_poweroff() can't confirm an abort request while it
waits for __pm_genpd_save_device() to return.
As this patch removes the intermediate states in pm_genpd_poweroff() while
powering off the PM domain, we no longer need the ability to abort that
sequence.
ii)
Make pm_runtime[_status]_suspended() reliable when used with genpd.
Until the last device in a genpd becomes idle, pm_genpd_runtime_suspend()
will return 0 without actually walking the hierarchy of the
->runtime_suspend() callbacks. However, by returning 0 the runtime PM core
considers the device as runtime_suspended, so
pm_runtime[_status]_suspended() will return true, even though the device
isn't (yet) runtime suspended.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks,
pm_runtime[_status]_suspended() will accurately reflect the status of the
device.
iii)
Enable fine-grained PM through runtime PM callbacks in drivers/subsystems.
There are currently cases were drivers/subsystems implements runtime PM
callbacks to deploy fine-grained PM (e.g. gate clocks, move pinctrl to
power-save state, etc.). While using the genpd, pm_genpd_runtime_suspend()
postpones invoking these callbacks until *all* the devices in the genpd
are runtime suspended. In essence, one runtime resumed device prevents
fine-grained PM for other devices within the same genpd.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks, fine-grained PM is enabled
throughout all the levels of runtime PM callbacks.
iiii)
Enable fine-grained PM for IRQ safe devices
Per the definition for an IRQ safe device, its runtime PM callbacks must
be able to execute in atomic context. In the path while genpd walks the
hierarchy of the ->runtime_suspend() callbacks for the device, it uses a
mutex. Therefore, genpd prevents that path to be executed for IRQ safe
devices.
As this patch changes pm_genpd_runtime_suspend() to immediately walk the
hierarchy of the ->runtime_suspend() callbacks and without needing to use
a mutex, fine-grained PM is enabled throughout all the levels of runtime
PM callbacks for IRQ safe devices.
Unfortunately this patch also comes with a drawback, as described in the
summary below.
Driver's/subsystem's runtime PM callbacks may be invoked even when the
genpd hasn't actually powered off the PM domain, potentially introducing
unnecessary latency.
However, in most cases, saving/restoring register contexts for devices are
typically fast operations or can be optimized in device specific ways
(e.g. shadow copies of register contents in memory, device-specific checks
to see if context has been lost before restoring context, etc.).
Still, in some cases the driver/subsystem may suffer from latency if
runtime PM is used in a very fine-grained manner (e.g. for each IO request
or xfer). To prevent that extra overhead, the driver/subsystem may deploy
the runtime PM autosuspend feature.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Reviewed-by: Kevin Hilman <khilman@linaro.org>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Tested-by: Lina Iyer <lina.iyer@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2015-06-18 13:17:53 +00:00
|
|
|
goto out;
|
2011-08-25 13:37:04 +00:00
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
PM: domains: Reverse the order of performance and enabling ops
The ->set_performance_state() needs to be called before ->power_on()
when a genpd is powered on, and after ->power_off() when a genpd is
powered off. Do this in order to let the provider know to which
performance state to power on the genpd, on the power on sequence, and
also to maintain the performance for that genpd until after powering off,
on power off sequence.
There is no scenario where a consumer would need its genpd enabled and
then its performance state increased. Instead, in every scenario, the
consumer needs the genpd to be enabled from the start at a specific
performance state.
And same logic applies to the powering down. No consumer would need its
genpd performance state dropped right before powering down.
Now, there are currently two vendors which use ->set_performance_state()
in their genpd providers. One of them is Tegra, but the only genpd provider
(PMC) that makes use of ->set_performance_state() doesn't implement the
->power_on() or ->power_off(), and so it will not be affected by the ops
reversal.
The other vendor that uses it is Qualcomm, in multiple genpd providers
actually (RPM, RPMh and CPR). But all Qualcomm genpd providers that make
use of ->set_performance_state() need the order between enabling ops and
the performance setting op to be reversed. And the reason for that is that
it currently translates into two different voltages in order to power on
a genpd to a specific performance state. Basically, ->power_on() switches
to the minimum (enabling) voltage for that genpd, and then
->set_performance_state() sets it to the voltage level required by the
consumer.
By reversing the call order, we rely on the provider to know what to do
on each call, but most popular usecase is to cache the performance state
and postpone the voltage setting until the ->power_on() gets called.
As for the reason of still needing the ->power_on() and ->power_off() for a
provider which could get away with just having ->set_performance_state()
implemented, there are consumers that do not (nor should) provide an
opp-table. For those consumers, ->set_performance_state() will not be
called, and so they will enable the genpd to its minimum performance state
by a ->power_on() call. Same logic goes for the disabling.
Signed-off-by: Abel Vesa <abel.vesa@linaro.org>
Reviewed-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2022-11-15 21:25:43 +00:00
|
|
|
genpd_restore_performance_state(dev, gpd_data->rpm_pstate);
|
2016-12-08 13:45:20 +00:00
|
|
|
ret = genpd_power_on(genpd, 0);
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
PM / Domains: Allow callbacks to execute all runtime PM helpers
A deadlock may occur if one of the PM domains' .start_device() or
.stop_device() callbacks or a device driver's .runtime_suspend() or
.runtime_resume() callback executed by the core generic PM domain
code uses a "wrong" runtime PM helper function. This happens, for
example, if .runtime_resume() from one device's driver calls
pm_runtime_resume() for another device in the same PM domain.
A similar situation may take place if a device's parent is in the
same PM domain, in which case the runtime PM framework may execute
pm_genpd_runtime_resume() automatically for the parent (if it is
suspended at the moment). This, of course, is undesirable, so
the generic PM domains code should be modified to prevent it from
happening.
The runtime PM framework guarantees that pm_genpd_runtime_suspend()
and pm_genpd_runtime_resume() won't be executed in parallel for
the same device, so the generic PM domains code need not worry
about those cases. Still, it needs to prevent the other possible
race conditions between pm_genpd_runtime_suspend(),
pm_genpd_runtime_resume(), pm_genpd_poweron() and pm_genpd_poweroff()
from happening and it needs to avoid deadlocks at the same time.
To this end, modify the generic PM domains code to relax
synchronization rules so that:
* pm_genpd_poweron() doesn't wait for the PM domain status to
change from GPD_STATE_BUSY. If it finds that the status is
not GPD_STATE_POWER_OFF, it returns without powering the domain on
(it may modify the status depending on the circumstances).
* pm_genpd_poweroff() returns as soon as it finds that the PM
domain's status changed from GPD_STATE_BUSY after it's released
the PM domain's lock.
* pm_genpd_runtime_suspend() doesn't wait for the PM domain status
to change from GPD_STATE_BUSY after executing the domain's
.stop_device() callback and executes pm_genpd_poweroff() only
if pm_genpd_runtime_resume() is not executed in parallel.
* pm_genpd_runtime_resume() doesn't wait for the PM domain status
to change from GPD_STATE_BUSY after executing pm_genpd_poweron()
and sets the domain's status to GPD_STATE_BUSY and increments its
counter of resuming devices (introduced by this change) immediately
after acquiring the lock. The counter of resuming devices is then
decremented after executing __pm_genpd_runtime_resume() for the
device and the domain's status is reset to GPD_STATE_ACTIVE (unless
there are more resuming devices in the domain, in which case the
status remains GPD_STATE_BUSY).
This way, for example, if a device driver's .runtime_resume()
callback executes pm_runtime_resume() for another device in the same
PM domain, pm_genpd_poweron() called by pm_genpd_runtime_resume()
invoked by the runtime PM framework will not block and it will see
that there's nothing to do for it. Next, the PM domain's lock will
be acquired without waiting for its status to change from
GPD_STATE_BUSY and the device driver's .runtime_resume() callback
will be executed. In turn, if pm_runtime_suspend() is executed by
one device driver's .runtime_resume() callback for another device in
the same PM domain, pm_genpd_poweroff() executed by
pm_genpd_runtime_suspend() invoked by the runtime PM framework as a
result will notice that one of the devices in the domain is being
resumed, so it will return immediately.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2011-07-11 22:39:36 +00:00
|
|
|
|
PM / Domains: Remove intermediate states from the power off sequence
Genpd's ->runtime_suspend() (assigned to pm_genpd_runtime_suspend())
doesn't immediately walk the hierarchy of ->runtime_suspend() callbacks.
Instead, pm_genpd_runtime_suspend() calls pm_genpd_poweroff() which
postpones that until *all* the devices in the genpd are runtime suspended.
When pm_genpd_poweroff() discovers that the last device in the genpd is
about to be runtime suspended, it calls __pm_genpd_save_device() for *all*
the devices in the genpd sequentially. Furthermore,
__pm_genpd_save_device() invokes the ->start() callback, walks the
hierarchy of the ->runtime_suspend() callbacks and invokes the ->stop()
callback. This causes a "thundering herd" problem.
Let's address this issue by having pm_genpd_runtime_suspend() immediately
walk the hierarchy of the ->runtime_suspend() callbacks, instead of
postponing that to the power off sequence via pm_genpd_poweroff(). If the
selected ->runtime_suspend() callback doesn't return an error code, call
pm_genpd_poweroff() to see if it's feasible to also power off the PM
domain.
Adopting this change enables us to simplify parts of the code in genpd,
for example the locking mechanism. Additionally, it gives some positive
side effects, as described below.
i)
One device's ->runtime_resume() latency is no longer affected by other
devices' latencies in a genpd.
The complexity genpd has to support the option to abort the power off
sequence suffers from latency issues. More precisely, a device that is
requested to be runtime resumed, may end up waiting for
__pm_genpd_save_device() to complete its operations for *another* device.
That's because pm_genpd_poweroff() can't confirm an abort request while it
waits for __pm_genpd_save_device() to return.
As this patch removes the intermediate states in pm_genpd_poweroff() while
powering off the PM domain, we no longer need the ability to abort that
sequence.
ii)
Make pm_runtime[_status]_suspended() reliable when used with genpd.
Until the last device in a genpd becomes idle, pm_genpd_runtime_suspend()
will return 0 without actually walking the hierarchy of the
->runtime_suspend() callbacks. However, by returning 0 the runtime PM core
considers the device as runtime_suspended, so
pm_runtime[_status]_suspended() will return true, even though the device
isn't (yet) runtime suspended.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks,
pm_runtime[_status]_suspended() will accurately reflect the status of the
device.
iii)
Enable fine-grained PM through runtime PM callbacks in drivers/subsystems.
There are currently cases were drivers/subsystems implements runtime PM
callbacks to deploy fine-grained PM (e.g. gate clocks, move pinctrl to
power-save state, etc.). While using the genpd, pm_genpd_runtime_suspend()
postpones invoking these callbacks until *all* the devices in the genpd
are runtime suspended. In essence, one runtime resumed device prevents
fine-grained PM for other devices within the same genpd.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks, fine-grained PM is enabled
throughout all the levels of runtime PM callbacks.
iiii)
Enable fine-grained PM for IRQ safe devices
Per the definition for an IRQ safe device, its runtime PM callbacks must
be able to execute in atomic context. In the path while genpd walks the
hierarchy of the ->runtime_suspend() callbacks for the device, it uses a
mutex. Therefore, genpd prevents that path to be executed for IRQ safe
devices.
As this patch changes pm_genpd_runtime_suspend() to immediately walk the
hierarchy of the ->runtime_suspend() callbacks and without needing to use
a mutex, fine-grained PM is enabled throughout all the levels of runtime
PM callbacks for IRQ safe devices.
Unfortunately this patch also comes with a drawback, as described in the
summary below.
Driver's/subsystem's runtime PM callbacks may be invoked even when the
genpd hasn't actually powered off the PM domain, potentially introducing
unnecessary latency.
However, in most cases, saving/restoring register contexts for devices are
typically fast operations or can be optimized in device specific ways
(e.g. shadow copies of register contents in memory, device-specific checks
to see if context has been lost before restoring context, etc.).
Still, in some cases the driver/subsystem may suffer from latency if
runtime PM is used in a very fine-grained manner (e.g. for each IO request
or xfer). To prevent that extra overhead, the driver/subsystem may deploy
the runtime PM autosuspend feature.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Reviewed-by: Kevin Hilman <khilman@linaro.org>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Tested-by: Lina Iyer <lina.iyer@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2015-06-18 13:17:53 +00:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
PM / Domains: Allow callbacks to execute all runtime PM helpers
A deadlock may occur if one of the PM domains' .start_device() or
.stop_device() callbacks or a device driver's .runtime_suspend() or
.runtime_resume() callback executed by the core generic PM domain
code uses a "wrong" runtime PM helper function. This happens, for
example, if .runtime_resume() from one device's driver calls
pm_runtime_resume() for another device in the same PM domain.
A similar situation may take place if a device's parent is in the
same PM domain, in which case the runtime PM framework may execute
pm_genpd_runtime_resume() automatically for the parent (if it is
suspended at the moment). This, of course, is undesirable, so
the generic PM domains code should be modified to prevent it from
happening.
The runtime PM framework guarantees that pm_genpd_runtime_suspend()
and pm_genpd_runtime_resume() won't be executed in parallel for
the same device, so the generic PM domains code need not worry
about those cases. Still, it needs to prevent the other possible
race conditions between pm_genpd_runtime_suspend(),
pm_genpd_runtime_resume(), pm_genpd_poweron() and pm_genpd_poweroff()
from happening and it needs to avoid deadlocks at the same time.
To this end, modify the generic PM domains code to relax
synchronization rules so that:
* pm_genpd_poweron() doesn't wait for the PM domain status to
change from GPD_STATE_BUSY. If it finds that the status is
not GPD_STATE_POWER_OFF, it returns without powering the domain on
(it may modify the status depending on the circumstances).
* pm_genpd_poweroff() returns as soon as it finds that the PM
domain's status changed from GPD_STATE_BUSY after it's released
the PM domain's lock.
* pm_genpd_runtime_suspend() doesn't wait for the PM domain status
to change from GPD_STATE_BUSY after executing the domain's
.stop_device() callback and executes pm_genpd_poweroff() only
if pm_genpd_runtime_resume() is not executed in parallel.
* pm_genpd_runtime_resume() doesn't wait for the PM domain status
to change from GPD_STATE_BUSY after executing pm_genpd_poweron()
and sets the domain's status to GPD_STATE_BUSY and increments its
counter of resuming devices (introduced by this change) immediately
after acquiring the lock. The counter of resuming devices is then
decremented after executing __pm_genpd_runtime_resume() for the
device and the domain's status is reset to GPD_STATE_ACTIVE (unless
there are more resuming devices in the domain, in which case the
status remains GPD_STATE_BUSY).
This way, for example, if a device driver's .runtime_resume()
callback executes pm_runtime_resume() for another device in the same
PM domain, pm_genpd_poweron() called by pm_genpd_runtime_resume()
invoked by the runtime PM framework will not block and it will see
that there's nothing to do for it. Next, the PM domain's lock will
be acquired without waiting for its status to change from
GPD_STATE_BUSY and the device driver's .runtime_resume() callback
will be executed. In turn, if pm_runtime_suspend() is executed by
one device driver's .runtime_resume() callback for another device in
the same PM domain, pm_genpd_poweroff() executed by
pm_genpd_runtime_suspend() invoked by the runtime PM framework as a
result will notice that one of the devices in the domain is being
resumed, so it will return immediately.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2011-07-11 22:39:36 +00:00
|
|
|
|
PM / Domains: Remove intermediate states from the power off sequence
Genpd's ->runtime_suspend() (assigned to pm_genpd_runtime_suspend())
doesn't immediately walk the hierarchy of ->runtime_suspend() callbacks.
Instead, pm_genpd_runtime_suspend() calls pm_genpd_poweroff() which
postpones that until *all* the devices in the genpd are runtime suspended.
When pm_genpd_poweroff() discovers that the last device in the genpd is
about to be runtime suspended, it calls __pm_genpd_save_device() for *all*
the devices in the genpd sequentially. Furthermore,
__pm_genpd_save_device() invokes the ->start() callback, walks the
hierarchy of the ->runtime_suspend() callbacks and invokes the ->stop()
callback. This causes a "thundering herd" problem.
Let's address this issue by having pm_genpd_runtime_suspend() immediately
walk the hierarchy of the ->runtime_suspend() callbacks, instead of
postponing that to the power off sequence via pm_genpd_poweroff(). If the
selected ->runtime_suspend() callback doesn't return an error code, call
pm_genpd_poweroff() to see if it's feasible to also power off the PM
domain.
Adopting this change enables us to simplify parts of the code in genpd,
for example the locking mechanism. Additionally, it gives some positive
side effects, as described below.
i)
One device's ->runtime_resume() latency is no longer affected by other
devices' latencies in a genpd.
The complexity genpd has to support the option to abort the power off
sequence suffers from latency issues. More precisely, a device that is
requested to be runtime resumed, may end up waiting for
__pm_genpd_save_device() to complete its operations for *another* device.
That's because pm_genpd_poweroff() can't confirm an abort request while it
waits for __pm_genpd_save_device() to return.
As this patch removes the intermediate states in pm_genpd_poweroff() while
powering off the PM domain, we no longer need the ability to abort that
sequence.
ii)
Make pm_runtime[_status]_suspended() reliable when used with genpd.
Until the last device in a genpd becomes idle, pm_genpd_runtime_suspend()
will return 0 without actually walking the hierarchy of the
->runtime_suspend() callbacks. However, by returning 0 the runtime PM core
considers the device as runtime_suspended, so
pm_runtime[_status]_suspended() will return true, even though the device
isn't (yet) runtime suspended.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks,
pm_runtime[_status]_suspended() will accurately reflect the status of the
device.
iii)
Enable fine-grained PM through runtime PM callbacks in drivers/subsystems.
There are currently cases were drivers/subsystems implements runtime PM
callbacks to deploy fine-grained PM (e.g. gate clocks, move pinctrl to
power-save state, etc.). While using the genpd, pm_genpd_runtime_suspend()
postpones invoking these callbacks until *all* the devices in the genpd
are runtime suspended. In essence, one runtime resumed device prevents
fine-grained PM for other devices within the same genpd.
After this patch, since pm_genpd_runtime_suspend() immediately walks the
hierarchy of the ->runtime_suspend() callbacks, fine-grained PM is enabled
throughout all the levels of runtime PM callbacks.
iiii)
Enable fine-grained PM for IRQ safe devices
Per the definition for an IRQ safe device, its runtime PM callbacks must
be able to execute in atomic context. In the path while genpd walks the
hierarchy of the ->runtime_suspend() callbacks for the device, it uses a
mutex. Therefore, genpd prevents that path to be executed for IRQ safe
devices.
As this patch changes pm_genpd_runtime_suspend() to immediately walk the
hierarchy of the ->runtime_suspend() callbacks and without needing to use
a mutex, fine-grained PM is enabled throughout all the levels of runtime
PM callbacks for IRQ safe devices.
Unfortunately this patch also comes with a drawback, as described in the
summary below.
Driver's/subsystem's runtime PM callbacks may be invoked even when the
genpd hasn't actually powered off the PM domain, potentially introducing
unnecessary latency.
However, in most cases, saving/restoring register contexts for devices are
typically fast operations or can be optimized in device specific ways
(e.g. shadow copies of register contents in memory, device-specific checks
to see if context has been lost before restoring context, etc.).
Still, in some cases the driver/subsystem may suffer from latency if
runtime PM is used in a very fine-grained manner (e.g. for each IO request
or xfer). To prevent that extra overhead, the driver/subsystem may deploy
the runtime PM autosuspend feature.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Reviewed-by: Kevin Hilman <khilman@linaro.org>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Tested-by: Lina Iyer <lina.iyer@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2015-06-18 13:17:53 +00:00
|
|
|
out:
|
2015-10-15 15:02:19 +00:00
|
|
|
/* Measure resume latency. */
|
2022-05-11 14:56:58 +00:00
|
|
|
if (timed)
|
2015-10-15 15:02:19 +00:00
|
|
|
time_start = ktime_get();
|
|
|
|
|
2016-03-01 23:20:38 +00:00
|
|
|
ret = genpd_start_dev(genpd, dev);
|
|
|
|
if (ret)
|
|
|
|
goto err_poweroff;
|
|
|
|
|
2016-03-31 09:21:27 +00:00
|
|
|
ret = __genpd_runtime_resume(dev);
|
2016-03-01 23:20:38 +00:00
|
|
|
if (ret)
|
|
|
|
goto err_stop;
|
2015-10-15 15:02:19 +00:00
|
|
|
|
|
|
|
/* Update resume latency value if the measured time exceeds it. */
|
2022-05-11 14:56:58 +00:00
|
|
|
if (timed) {
|
2015-10-15 15:02:19 +00:00
|
|
|
elapsed_ns = ktime_to_ns(ktime_sub(ktime_get(), time_start));
|
2022-05-11 14:56:58 +00:00
|
|
|
if (elapsed_ns > td->resume_latency_ns) {
|
2015-10-15 15:02:19 +00:00
|
|
|
td->resume_latency_ns = elapsed_ns;
|
|
|
|
dev_dbg(dev, "resume latency exceeded, %lld ns\n",
|
|
|
|
elapsed_ns);
|
2022-05-11 14:57:02 +00:00
|
|
|
genpd->gd->max_off_time_changed = true;
|
2015-10-15 15:02:19 +00:00
|
|
|
td->constraint_changed = true;
|
|
|
|
}
|
|
|
|
}
|
2011-07-11 22:39:29 +00:00
|
|
|
|
2011-07-01 20:12:45 +00:00
|
|
|
return 0;
|
2016-03-01 23:20:38 +00:00
|
|
|
|
|
|
|
err_stop:
|
|
|
|
genpd_stop_dev(genpd, dev);
|
|
|
|
err_poweroff:
|
2021-01-27 08:42:05 +00:00
|
|
|
if (!pm_runtime_is_irq_safe(dev) || genpd_is_irq_safe(genpd)) {
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
2017-02-17 09:55:25 +00:00
|
|
|
genpd_power_off(genpd, true, 0);
|
PM: domains: Reverse the order of performance and enabling ops
The ->set_performance_state() needs to be called before ->power_on()
when a genpd is powered on, and after ->power_off() when a genpd is
powered off. Do this in order to let the provider know to which
performance state to power on the genpd, on the power on sequence, and
also to maintain the performance for that genpd until after powering off,
on power off sequence.
There is no scenario where a consumer would need its genpd enabled and
then its performance state increased. Instead, in every scenario, the
consumer needs the genpd to be enabled from the start at a specific
performance state.
And same logic applies to the powering down. No consumer would need its
genpd performance state dropped right before powering down.
Now, there are currently two vendors which use ->set_performance_state()
in their genpd providers. One of them is Tegra, but the only genpd provider
(PMC) that makes use of ->set_performance_state() doesn't implement the
->power_on() or ->power_off(), and so it will not be affected by the ops
reversal.
The other vendor that uses it is Qualcomm, in multiple genpd providers
actually (RPM, RPMh and CPR). But all Qualcomm genpd providers that make
use of ->set_performance_state() need the order between enabling ops and
the performance setting op to be reversed. And the reason for that is that
it currently translates into two different voltages in order to power on
a genpd to a specific performance state. Basically, ->power_on() switches
to the minimum (enabling) voltage for that genpd, and then
->set_performance_state() sets it to the voltage level required by the
consumer.
By reversing the call order, we rely on the provider to know what to do
on each call, but most popular usecase is to cache the performance state
and postpone the voltage setting until the ->power_on() gets called.
As for the reason of still needing the ->power_on() and ->power_off() for a
provider which could get away with just having ->set_performance_state()
implemented, there are consumers that do not (nor should) provide an
opp-table. For those consumers, ->set_performance_state() will not be
called, and so they will enable the genpd to its minimum performance state
by a ->power_on() call. Same logic goes for the disabling.
Signed-off-by: Abel Vesa <abel.vesa@linaro.org>
Reviewed-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2022-11-15 21:25:43 +00:00
|
|
|
gpd_data->rpm_pstate = genpd_drop_performance_state(dev);
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2016-03-01 23:20:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
2011-07-01 20:12:45 +00:00
|
|
|
}
|
|
|
|
|
2014-03-28 05:20:21 +00:00
|
|
|
static bool pd_ignore_unused;
|
|
|
|
static int __init pd_ignore_unused_setup(char *__unused)
|
|
|
|
{
|
|
|
|
pd_ignore_unused = true;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
__setup("pd_ignore_unused", pd_ignore_unused_setup);
|
|
|
|
|
2011-08-14 11:34:31 +00:00
|
|
|
/**
|
2016-12-08 13:45:20 +00:00
|
|
|
* genpd_power_off_unused - Power off all PM domains with no devices in use.
|
2011-08-14 11:34:31 +00:00
|
|
|
*/
|
2016-12-08 13:45:20 +00:00
|
|
|
static int __init genpd_power_off_unused(void)
|
2011-08-14 11:34:31 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
|
|
|
|
2014-03-28 05:20:21 +00:00
|
|
|
if (pd_ignore_unused) {
|
|
|
|
pr_warn("genpd: Not disabling unused power domains\n");
|
2015-10-06 12:27:42 +00:00
|
|
|
return 0;
|
2014-03-28 05:20:21 +00:00
|
|
|
}
|
|
|
|
|
2023-12-27 15:18:54 +00:00
|
|
|
pr_info("genpd: Disabling unused power domains\n");
|
2011-08-14 11:34:31 +00:00
|
|
|
mutex_lock(&gpd_list_lock);
|
|
|
|
|
|
|
|
list_for_each_entry(genpd, &gpd_list, gpd_list_node)
|
|
|
|
genpd_queue_power_off_work(genpd);
|
|
|
|
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
2014-09-03 10:52:26 +00:00
|
|
|
return 0;
|
|
|
|
}
|
2023-12-27 15:21:24 +00:00
|
|
|
late_initcall_sync(genpd_power_off_unused);
|
2014-09-03 10:52:26 +00:00
|
|
|
|
2016-09-12 11:01:10 +00:00
|
|
|
#ifdef CONFIG_PM_SLEEP
|
|
|
|
|
2011-07-01 20:13:19 +00:00
|
|
|
/**
|
2020-07-08 23:32:13 +00:00
|
|
|
* genpd_sync_power_off - Synchronously power off a PM domain and its parents.
|
2011-07-01 20:13:19 +00:00
|
|
|
* @genpd: PM domain to power off, if possible.
|
2017-02-08 12:39:00 +00:00
|
|
|
* @use_lock: use the lock.
|
|
|
|
* @depth: nesting count for lockdep.
|
2011-07-01 20:13:19 +00:00
|
|
|
*
|
|
|
|
* Check if the given PM domain can be powered off (during system suspend or
|
2020-07-08 23:32:13 +00:00
|
|
|
* hibernation) and do that if so. Also, in that case propagate to its parents.
|
2011-07-01 20:13:19 +00:00
|
|
|
*
|
2012-08-05 23:39:57 +00:00
|
|
|
* This function is only called in "noirq" and "syscore" stages of system power
|
2017-02-08 12:39:00 +00:00
|
|
|
* transitions. The "noirq" callbacks may be executed asynchronously, thus in
|
|
|
|
* these cases the lock must be held.
|
2011-07-01 20:13:19 +00:00
|
|
|
*/
|
2017-02-08 12:39:00 +00:00
|
|
|
static void genpd_sync_power_off(struct generic_pm_domain *genpd, bool use_lock,
|
|
|
|
unsigned int depth)
|
2011-07-01 20:13:19 +00:00
|
|
|
{
|
2011-08-08 21:43:40 +00:00
|
|
|
struct gpd_link *link;
|
2011-07-01 20:13:19 +00:00
|
|
|
|
2017-03-20 10:19:21 +00:00
|
|
|
if (!genpd_status_on(genpd) || genpd_is_always_on(genpd))
|
2011-07-01 20:13:19 +00:00
|
|
|
return;
|
|
|
|
|
2011-08-08 21:43:04 +00:00
|
|
|
if (genpd->suspended_count != genpd->device_count
|
|
|
|
|| atomic_read(&genpd->sd_count) > 0)
|
2011-07-01 20:13:19 +00:00
|
|
|
return;
|
|
|
|
|
2022-02-17 12:49:50 +00:00
|
|
|
/* Check that the children are in their deepest (powered-off) state. */
|
|
|
|
list_for_each_entry(link, &genpd->parent_links, parent_node) {
|
|
|
|
struct generic_pm_domain *child = link->child;
|
|
|
|
if (child->state_idx < child->state_count - 1)
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2016-02-15 10:10:51 +00:00
|
|
|
/* Choose the deepest state when suspending */
|
|
|
|
genpd->state_idx = genpd->state_count - 1;
|
2024-04-12 10:42:07 +00:00
|
|
|
if (_genpd_power_off(genpd, false)) {
|
|
|
|
genpd->states[genpd->state_idx].rejected++;
|
2017-03-20 10:19:22 +00:00
|
|
|
return;
|
2024-04-12 10:42:07 +00:00
|
|
|
} else {
|
|
|
|
genpd->states[genpd->state_idx].usage++;
|
|
|
|
}
|
2011-07-01 20:13:19 +00:00
|
|
|
|
2020-09-24 11:04:47 +00:00
|
|
|
genpd->status = GENPD_STATE_OFF;
|
2011-08-08 21:43:40 +00:00
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry(link, &genpd->child_links, child_node) {
|
|
|
|
genpd_sd_counter_dec(link->parent);
|
2017-02-08 12:39:00 +00:00
|
|
|
|
|
|
|
if (use_lock)
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_lock_nested(link->parent, depth + 1);
|
2017-02-08 12:39:00 +00:00
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_sync_power_off(link->parent, use_lock, depth + 1);
|
2017-02-08 12:39:00 +00:00
|
|
|
|
|
|
|
if (use_lock)
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_unlock(link->parent);
|
2011-07-01 20:13:19 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-08-05 23:39:16 +00:00
|
|
|
/**
|
2020-07-08 23:32:13 +00:00
|
|
|
* genpd_sync_power_on - Synchronously power on a PM domain and its parents.
|
2012-08-05 23:39:16 +00:00
|
|
|
* @genpd: PM domain to power on.
|
2017-02-08 12:39:00 +00:00
|
|
|
* @use_lock: use the lock.
|
|
|
|
* @depth: nesting count for lockdep.
|
2012-08-05 23:39:16 +00:00
|
|
|
*
|
2012-08-05 23:39:57 +00:00
|
|
|
* This function is only called in "noirq" and "syscore" stages of system power
|
2017-02-08 12:39:00 +00:00
|
|
|
* transitions. The "noirq" callbacks may be executed asynchronously, thus in
|
|
|
|
* these cases the lock must be held.
|
2012-08-05 23:39:16 +00:00
|
|
|
*/
|
2017-02-08 12:39:00 +00:00
|
|
|
static void genpd_sync_power_on(struct generic_pm_domain *genpd, bool use_lock,
|
|
|
|
unsigned int depth)
|
2012-08-05 23:39:16 +00:00
|
|
|
{
|
|
|
|
struct gpd_link *link;
|
|
|
|
|
2017-03-20 10:19:20 +00:00
|
|
|
if (genpd_status_on(genpd))
|
2012-08-05 23:39:16 +00:00
|
|
|
return;
|
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry(link, &genpd->child_links, child_node) {
|
|
|
|
genpd_sd_counter_inc(link->parent);
|
2017-02-08 12:39:00 +00:00
|
|
|
|
|
|
|
if (use_lock)
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_lock_nested(link->parent, depth + 1);
|
2017-02-08 12:39:00 +00:00
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_sync_power_on(link->parent, use_lock, depth + 1);
|
2017-02-08 12:39:00 +00:00
|
|
|
|
|
|
|
if (use_lock)
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_unlock(link->parent);
|
2012-08-05 23:39:16 +00:00
|
|
|
}
|
|
|
|
|
2016-12-08 13:45:20 +00:00
|
|
|
_genpd_power_on(genpd, false);
|
2020-09-24 11:04:47 +00:00
|
|
|
genpd->status = GENPD_STATE_ON;
|
2012-08-05 23:39:16 +00:00
|
|
|
}
|
|
|
|
|
2011-07-01 20:13:19 +00:00
|
|
|
/**
|
2017-10-06 07:02:06 +00:00
|
|
|
* genpd_prepare - Start power transition of a device in a PM domain.
|
2011-07-01 20:13:19 +00:00
|
|
|
* @dev: Device to start the transition of.
|
|
|
|
*
|
|
|
|
* Start a power transition of a device (during a system-wide power transition)
|
|
|
|
* under the assumption that its pm_domain field points to the domain member of
|
|
|
|
* an object of type struct generic_pm_domain representing a PM domain
|
|
|
|
* consisting of I/O devices.
|
|
|
|
*/
|
2017-10-06 07:02:06 +00:00
|
|
|
static int genpd_prepare(struct device *dev)
|
2011-07-01 20:13:19 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
2011-07-11 22:39:21 +00:00
|
|
|
int ret;
|
2011-07-01 20:13:19 +00:00
|
|
|
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
|
|
|
genpd = dev_to_genpd(dev);
|
|
|
|
if (IS_ERR(genpd))
|
|
|
|
return -EINVAL;
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
2024-04-12 10:42:08 +00:00
|
|
|
genpd->prepared_count++;
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2011-07-01 20:13:19 +00:00
|
|
|
|
2011-07-11 22:39:21 +00:00
|
|
|
ret = pm_generic_prepare(dev);
|
PM / Domains: Fix genpd to deal with drivers returning 1 from ->prepare()
During system-wide PM, genpd relies on its PM callbacks to be invoked for
all its attached devices, as to deal with powering off/on the PM domain. In
other words, genpd is not compatible with the direct_complete path, if
executed by the PM core for any of its attached devices.
However, when genpd's ->prepare() callback invokes pm_generic_prepare(), it
does not take into account that it may return 1. Instead it treats that as
an error internally and expects the PM core to abort the prepare phase and
roll back. This leads to genpd not properly powering on/off the PM domain,
because its internal counters gets wrongly balanced.
To fix the behaviour, allow drivers to return 1 from their ->prepare()
callbacks, but let's return 0 from genpd's ->prepare() callback in such
case, as that prevents the PM core from running the direct_complete path
for the device.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2017-11-08 09:11:02 +00:00
|
|
|
if (ret < 0) {
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
2011-07-11 22:39:21 +00:00
|
|
|
|
PM / Domains: Allow genpd to power on during system PM phases
If a PM domain is powered off when the first device starts its system PM
prepare phase, genpd prevents any further attempts to power on the PM
domain during the following system PM phases. Not until the system PM
complete phase is finalized for all devices in the PM domain, genpd again
allows it to be powered on.
This behaviour needs to be changed, as a subsystem/driver for a device in
the same PM domain may still need to be able to serve requests in some of
the system PM phases. Accordingly, it may need to runtime resume its
device and thus also request the corresponding PM domain to be powered on.
To deal with these scenarios, let's make the device operational in the
system PM prepare phase by runtime resuming it, no matter if the PM domain
is powered on or off. Changing this also enables us to remove genpd's
suspend_power_off flag, as it's being used to track this condition.
Additionally, we must allow the PM domain to be powered on via runtime PM
during the system PM phases.
This change also requires a fix in the AMD ACP (Audio CoProcessor) drm
driver. It registers a genpd to model the ACP as a PM domain, but
unfortunately it's also abuses genpd's "internal" suspend_power_off flag
to deal with a corner case at system PM resume.
More precisely, the so called SMU block powers on the ACP at system PM
resume, unconditionally if it's being used or not. This may lead to that
genpd's internal status of the power state, may not correctly reflect the
power state of the HW after a system PM resume.
Because of changing the behaviour of genpd, by runtime resuming devices in
the prepare phase, the AMD ACP drm driver no longer have to deal with this
corner case. So let's just drop the related code in this driver.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Reviewed-by: Kevin Hilman <khilman@baylibre.com>
Acked-by: Maruthi Bayyavarapu <maruthi.bayyavarapu@amd.com>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2016-05-30 09:43:07 +00:00
|
|
|
genpd->prepared_count--;
|
2011-07-11 22:39:21 +00:00
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2011-07-11 22:39:21 +00:00
|
|
|
}
|
2011-07-11 22:39:29 +00:00
|
|
|
|
PM / Domains: Fix genpd to deal with drivers returning 1 from ->prepare()
During system-wide PM, genpd relies on its PM callbacks to be invoked for
all its attached devices, as to deal with powering off/on the PM domain. In
other words, genpd is not compatible with the direct_complete path, if
executed by the PM core for any of its attached devices.
However, when genpd's ->prepare() callback invokes pm_generic_prepare(), it
does not take into account that it may return 1. Instead it treats that as
an error internally and expects the PM core to abort the prepare phase and
roll back. This leads to genpd not properly powering on/off the PM domain,
because its internal counters gets wrongly balanced.
To fix the behaviour, allow drivers to return 1 from their ->prepare()
callbacks, but let's return 0 from genpd's ->prepare() callback in such
case, as that prevents the PM core from running the direct_complete path
for the device.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2017-11-08 09:11:02 +00:00
|
|
|
/* Never return 1, as genpd don't cope with the direct_complete path. */
|
|
|
|
return ret >= 0 ? 0 : ret;
|
2011-07-01 20:13:19 +00:00
|
|
|
}
|
|
|
|
|
2012-01-29 19:39:02 +00:00
|
|
|
/**
|
2017-06-22 07:18:33 +00:00
|
|
|
* genpd_finish_suspend - Completion of suspend or hibernation of device in an
|
|
|
|
* I/O pm domain.
|
2012-01-29 19:39:02 +00:00
|
|
|
* @dev: Device to suspend.
|
2022-11-02 14:21:02 +00:00
|
|
|
* @suspend_noirq: Generic suspend_noirq callback.
|
|
|
|
* @resume_noirq: Generic resume_noirq callback.
|
2012-01-29 19:39:02 +00:00
|
|
|
*
|
|
|
|
* Stop the device and remove power from the domain if all devices in it have
|
|
|
|
* been stopped.
|
|
|
|
*/
|
2022-11-02 14:21:02 +00:00
|
|
|
static int genpd_finish_suspend(struct device *dev,
|
|
|
|
int (*suspend_noirq)(struct device *dev),
|
|
|
|
int (*resume_noirq)(struct device *dev))
|
2012-01-29 19:39:02 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
2018-01-10 20:31:56 +00:00
|
|
|
int ret = 0;
|
2012-01-29 19:39:02 +00:00
|
|
|
|
|
|
|
genpd = dev_to_genpd(dev);
|
|
|
|
if (IS_ERR(genpd))
|
|
|
|
return -EINVAL;
|
2011-07-01 20:13:19 +00:00
|
|
|
|
2022-11-02 14:21:02 +00:00
|
|
|
ret = suspend_noirq(dev);
|
2017-06-22 07:18:33 +00:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
2020-11-19 07:25:39 +00:00
|
|
|
if (device_wakeup_path(dev) && genpd_is_active_wakeup(genpd))
|
2018-01-10 20:31:56 +00:00
|
|
|
return 0;
|
|
|
|
|
PM / genpd: Stop/start devices without pm_runtime_force_suspend/resume()
There are problems with calling pm_runtime_force_suspend/resume()
to "stop" and "start" devices in genpd_finish_suspend() and
genpd_resume_noirq() (and in analogous hibernation-specific genpd
callbacks) after commit 122a22377a3d (PM / Domains: Stop/start
devices during system PM suspend/resume in genpd) as those routines
do much more than just "stopping" and "starting" devices (which was
the stated purpose of that commit) unnecessarily and may not play
well with system-wide PM driver callbacks.
First, consider the pm_runtime_force_suspend() in
genpd_finish_suspend(). If the current runtime PM status of the
device is "suspended", that function most likely does the right thing
by ignoring the device, because it should have been "stopped" already
and whatever needed to be done to deactivate it shoud have been done.
In turn, if the runtime PM status of the device is "active",
genpd_runtime_suspend() is called for it (indirectly) and (1) runs
the ->runtime_suspend callback provided by the device's driver
(assuming no bus type with ->runtime_suspend of its own), (2) "stops"
the device and (3) checks if the domain can be powered down, and then
(4) the device's runtime PM status is changed to "suspended". Out of
the four actions above (1) is not necessary and it may be outright
harmful, (3) is pointless and (4) is questionable. The only
operation that needs to be carried out here is (2).
The reason why (1) is not necessary is because the system-wide
PM callbacks provided by the device driver for the transition in
question have been run and they should have taken care of the
driver's part of device suspend already. Moreover, it may be
harmful, because the ->runtime_suspend callback may want to
access the device which is partially suspended at that point
and may not be responsive. Also, system-wide PM callbacks may
have been run already (in the previous phases of the system
transition under way) for the device's parent or for its supplier
devices (if any) and the device may not be accessible because of
that.
There also is no reason to do (3), because genpd_finish_suspend()
will repeat it anyway, and (4) potentially causes confusion to ensue
during the subsequent system transition to the working state.
Consider pm_runtime_force_resume() in genpd_resume_noirq() now.
It runs genpd_runtime_resume() for all devices with runtime PM
status set to "suspended", which includes all of the devices
whose runtime PM status was changed by pm_runtime_force_suspend()
before and may include some devices already suspended when the
pm_runtime_force_suspend() was running, which may be confusing. The
genpd_runtime_resume() first tries to power up the domain, which
(again) is pointless, because genpd_resume_noirq() has done that
already. Then, it "starts" the device and runs the ->runtime_resume
callback (from the driver, say) for it. If all is well, the device
is left with the runtime PM status set to "active".
Unfortunately, running the driver's ->runtime_resume callback
before its system-wide PM callbacks and possibly before some
system-wide PM callbacks of the parent device's driver (let
alone supplier drivers) is asking for trouble, especially if
the device had been suspended before pm_runtime_force_suspend()
ran previously or if the callbacks in question expect to be run
back-to-back with their suspend-side counterparts. It also should
not be necessary, because the system-wide PM driver callbacks that
will be invoked for the device subsequently should take care of
resuming it just fine.
[Running the driver's ->runtime_resume callback in the "noirq"
phase of the transition to the working state may be problematic
even for devices whose drivers do use pm_runtime_force_resume()
in (or as) their system-wide PM callbacks if they have suppliers
other than their parents, because it may cause the supplier to
be resumed after the consumer in some cases.]
Because of the above, modify genpd as follows:
1. Change genpd_finish_suspend() to only "stop" devices with
runtime PM status set to "active" (without invoking runtime PM
callbacks for them, changing their runtime PM status and so on).
That doesn't change the handling of devices whose drivers use
pm_runtime_force_suspend/resume() in (or as) their system-wide
PM callbacks and addresses the issues described above for the
other devices.
2. Change genpd_resume_noirq() to only "start" devices with
runtime PM status set to "active" (without invoking runtime PM
callbacks for them, changing their runtime PM status and so on).
Again, that doesn't change the handling of devices whose drivers
use pm_runtime_force_suspend/resume() in (or as) their system-wide
PM callbacks and addresses the described issues for the other
devices. Devices with runtime PM status set to "suspended"
are not started with the assumption that they will be resumed
later, either by pm_runtime_force_resume() or via runtime PM.
3. Change genpd_restore_noirq() to follow genpd_resume_noirq().
That causes devices already suspended before hibernation to be
left alone (which also is the case without the change) and
avoids running the ->runtime_resume driver callback too early
for the other devices.
4. Change genpd_freeze_noirq() and genpd_thaw_noirq() in accordance
with the above modifications.
Fixes: 122a22377a3d (PM / Domains: Stop/start devices during system PM suspend/resume in genpd)
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
Acked-by: Ulf Hansson <ulf.hansson@linaro.org>
2018-01-12 13:10:38 +00:00
|
|
|
if (genpd->dev_ops.stop && genpd->dev_ops.start &&
|
|
|
|
!pm_runtime_status_suspended(dev)) {
|
|
|
|
ret = genpd_stop_dev(genpd, dev);
|
2018-01-10 20:31:56 +00:00
|
|
|
if (ret) {
|
2022-11-02 14:21:02 +00:00
|
|
|
resume_noirq(dev);
|
2016-05-30 09:33:14 +00:00
|
|
|
return ret;
|
2018-01-10 20:31:56 +00:00
|
|
|
}
|
2016-05-30 09:33:14 +00:00
|
|
|
}
|
|
|
|
|
2017-02-08 12:39:00 +00:00
|
|
|
genpd_lock(genpd);
|
2011-07-01 20:13:19 +00:00
|
|
|
genpd->suspended_count++;
|
2017-02-08 12:39:00 +00:00
|
|
|
genpd_sync_power_off(genpd, true, 0);
|
|
|
|
genpd_unlock(genpd);
|
2011-07-01 20:13:19 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-06-22 07:18:33 +00:00
|
|
|
/**
|
2017-10-06 07:02:06 +00:00
|
|
|
* genpd_suspend_noirq - Completion of suspend of device in an I/O PM domain.
|
2017-06-22 07:18:33 +00:00
|
|
|
* @dev: Device to suspend.
|
|
|
|
*
|
|
|
|
* Stop the device and remove power from the domain if all devices in it have
|
|
|
|
* been stopped.
|
|
|
|
*/
|
2017-10-06 07:02:06 +00:00
|
|
|
static int genpd_suspend_noirq(struct device *dev)
|
2017-06-22 07:18:33 +00:00
|
|
|
{
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
2022-11-02 14:21:02 +00:00
|
|
|
return genpd_finish_suspend(dev,
|
|
|
|
pm_generic_suspend_noirq,
|
|
|
|
pm_generic_resume_noirq);
|
2017-06-22 07:18:33 +00:00
|
|
|
}
|
|
|
|
|
2011-07-01 20:13:19 +00:00
|
|
|
/**
|
2022-11-02 14:21:03 +00:00
|
|
|
* genpd_finish_resume - Completion of resume of device in an I/O PM domain.
|
2011-07-01 20:13:19 +00:00
|
|
|
* @dev: Device to resume.
|
2022-11-02 14:21:03 +00:00
|
|
|
* @resume_noirq: Generic resume_noirq callback.
|
2011-07-01 20:13:19 +00:00
|
|
|
*
|
2012-01-29 19:39:02 +00:00
|
|
|
* Restore power to the device's PM domain, if necessary, and start the device.
|
2011-07-01 20:13:19 +00:00
|
|
|
*/
|
2022-11-02 14:21:03 +00:00
|
|
|
static int genpd_finish_resume(struct device *dev,
|
|
|
|
int (*resume_noirq)(struct device *dev))
|
2011-07-01 20:13:19 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
2018-01-10 20:31:56 +00:00
|
|
|
int ret;
|
2011-07-01 20:13:19 +00:00
|
|
|
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
|
|
|
genpd = dev_to_genpd(dev);
|
|
|
|
if (IS_ERR(genpd))
|
|
|
|
return -EINVAL;
|
|
|
|
|
2020-11-19 07:25:39 +00:00
|
|
|
if (device_wakeup_path(dev) && genpd_is_active_wakeup(genpd))
|
2022-11-02 14:21:03 +00:00
|
|
|
return resume_noirq(dev);
|
2011-07-01 20:13:19 +00:00
|
|
|
|
2017-02-08 12:39:00 +00:00
|
|
|
genpd_lock(genpd);
|
|
|
|
genpd_sync_power_on(genpd, true, 0);
|
2011-07-01 20:13:19 +00:00
|
|
|
genpd->suspended_count--;
|
2017-02-08 12:39:00 +00:00
|
|
|
genpd_unlock(genpd);
|
2011-07-01 20:13:19 +00:00
|
|
|
|
PM / genpd: Stop/start devices without pm_runtime_force_suspend/resume()
There are problems with calling pm_runtime_force_suspend/resume()
to "stop" and "start" devices in genpd_finish_suspend() and
genpd_resume_noirq() (and in analogous hibernation-specific genpd
callbacks) after commit 122a22377a3d (PM / Domains: Stop/start
devices during system PM suspend/resume in genpd) as those routines
do much more than just "stopping" and "starting" devices (which was
the stated purpose of that commit) unnecessarily and may not play
well with system-wide PM driver callbacks.
First, consider the pm_runtime_force_suspend() in
genpd_finish_suspend(). If the current runtime PM status of the
device is "suspended", that function most likely does the right thing
by ignoring the device, because it should have been "stopped" already
and whatever needed to be done to deactivate it shoud have been done.
In turn, if the runtime PM status of the device is "active",
genpd_runtime_suspend() is called for it (indirectly) and (1) runs
the ->runtime_suspend callback provided by the device's driver
(assuming no bus type with ->runtime_suspend of its own), (2) "stops"
the device and (3) checks if the domain can be powered down, and then
(4) the device's runtime PM status is changed to "suspended". Out of
the four actions above (1) is not necessary and it may be outright
harmful, (3) is pointless and (4) is questionable. The only
operation that needs to be carried out here is (2).
The reason why (1) is not necessary is because the system-wide
PM callbacks provided by the device driver for the transition in
question have been run and they should have taken care of the
driver's part of device suspend already. Moreover, it may be
harmful, because the ->runtime_suspend callback may want to
access the device which is partially suspended at that point
and may not be responsive. Also, system-wide PM callbacks may
have been run already (in the previous phases of the system
transition under way) for the device's parent or for its supplier
devices (if any) and the device may not be accessible because of
that.
There also is no reason to do (3), because genpd_finish_suspend()
will repeat it anyway, and (4) potentially causes confusion to ensue
during the subsequent system transition to the working state.
Consider pm_runtime_force_resume() in genpd_resume_noirq() now.
It runs genpd_runtime_resume() for all devices with runtime PM
status set to "suspended", which includes all of the devices
whose runtime PM status was changed by pm_runtime_force_suspend()
before and may include some devices already suspended when the
pm_runtime_force_suspend() was running, which may be confusing. The
genpd_runtime_resume() first tries to power up the domain, which
(again) is pointless, because genpd_resume_noirq() has done that
already. Then, it "starts" the device and runs the ->runtime_resume
callback (from the driver, say) for it. If all is well, the device
is left with the runtime PM status set to "active".
Unfortunately, running the driver's ->runtime_resume callback
before its system-wide PM callbacks and possibly before some
system-wide PM callbacks of the parent device's driver (let
alone supplier drivers) is asking for trouble, especially if
the device had been suspended before pm_runtime_force_suspend()
ran previously or if the callbacks in question expect to be run
back-to-back with their suspend-side counterparts. It also should
not be necessary, because the system-wide PM driver callbacks that
will be invoked for the device subsequently should take care of
resuming it just fine.
[Running the driver's ->runtime_resume callback in the "noirq"
phase of the transition to the working state may be problematic
even for devices whose drivers do use pm_runtime_force_resume()
in (or as) their system-wide PM callbacks if they have suppliers
other than their parents, because it may cause the supplier to
be resumed after the consumer in some cases.]
Because of the above, modify genpd as follows:
1. Change genpd_finish_suspend() to only "stop" devices with
runtime PM status set to "active" (without invoking runtime PM
callbacks for them, changing their runtime PM status and so on).
That doesn't change the handling of devices whose drivers use
pm_runtime_force_suspend/resume() in (or as) their system-wide
PM callbacks and addresses the issues described above for the
other devices.
2. Change genpd_resume_noirq() to only "start" devices with
runtime PM status set to "active" (without invoking runtime PM
callbacks for them, changing their runtime PM status and so on).
Again, that doesn't change the handling of devices whose drivers
use pm_runtime_force_suspend/resume() in (or as) their system-wide
PM callbacks and addresses the described issues for the other
devices. Devices with runtime PM status set to "suspended"
are not started with the assumption that they will be resumed
later, either by pm_runtime_force_resume() or via runtime PM.
3. Change genpd_restore_noirq() to follow genpd_resume_noirq().
That causes devices already suspended before hibernation to be
left alone (which also is the case without the change) and
avoids running the ->runtime_resume driver callback too early
for the other devices.
4. Change genpd_freeze_noirq() and genpd_thaw_noirq() in accordance
with the above modifications.
Fixes: 122a22377a3d (PM / Domains: Stop/start devices during system PM suspend/resume in genpd)
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
Acked-by: Ulf Hansson <ulf.hansson@linaro.org>
2018-01-12 13:10:38 +00:00
|
|
|
if (genpd->dev_ops.stop && genpd->dev_ops.start &&
|
|
|
|
!pm_runtime_status_suspended(dev)) {
|
|
|
|
ret = genpd_start_dev(genpd, dev);
|
2018-01-10 20:31:56 +00:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
}
|
2016-05-30 09:33:14 +00:00
|
|
|
|
2018-01-10 20:31:56 +00:00
|
|
|
return pm_generic_resume_noirq(dev);
|
2011-07-01 20:13:19 +00:00
|
|
|
}
|
|
|
|
|
2022-11-02 14:21:03 +00:00
|
|
|
/**
|
|
|
|
* genpd_resume_noirq - Start of resume of device in an I/O PM domain.
|
|
|
|
* @dev: Device to resume.
|
|
|
|
*
|
|
|
|
* Restore power to the device's PM domain, if necessary, and start the device.
|
|
|
|
*/
|
|
|
|
static int genpd_resume_noirq(struct device *dev)
|
|
|
|
{
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
|
|
|
return genpd_finish_resume(dev, pm_generic_resume_noirq);
|
|
|
|
}
|
|
|
|
|
2012-01-29 19:39:02 +00:00
|
|
|
/**
|
2017-10-06 07:02:06 +00:00
|
|
|
* genpd_freeze_noirq - Completion of freezing a device in an I/O PM domain.
|
2011-07-01 20:13:19 +00:00
|
|
|
* @dev: Device to freeze.
|
|
|
|
*
|
|
|
|
* Carry out a late freeze of a device under the assumption that its
|
|
|
|
* pm_domain field points to the domain member of an object of type
|
|
|
|
* struct generic_pm_domain representing a power domain consisting of I/O
|
|
|
|
* devices.
|
|
|
|
*/
|
2017-10-06 07:02:06 +00:00
|
|
|
static int genpd_freeze_noirq(struct device *dev)
|
2011-07-01 20:13:19 +00:00
|
|
|
{
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
2022-11-02 14:21:04 +00:00
|
|
|
return genpd_finish_suspend(dev,
|
|
|
|
pm_generic_freeze_noirq,
|
|
|
|
pm_generic_thaw_noirq);
|
2012-01-29 19:39:02 +00:00
|
|
|
}
|
2011-07-01 20:13:19 +00:00
|
|
|
|
2012-01-29 19:39:02 +00:00
|
|
|
/**
|
2017-10-06 07:02:06 +00:00
|
|
|
* genpd_thaw_noirq - Early thaw of device in an I/O PM domain.
|
2012-01-29 19:39:02 +00:00
|
|
|
* @dev: Device to thaw.
|
|
|
|
*
|
|
|
|
* Start the device, unless power has been removed from the domain already
|
|
|
|
* before the system transition.
|
|
|
|
*/
|
2017-10-06 07:02:06 +00:00
|
|
|
static int genpd_thaw_noirq(struct device *dev)
|
2012-01-29 19:39:02 +00:00
|
|
|
{
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
2011-07-01 20:13:19 +00:00
|
|
|
|
2022-11-02 14:21:04 +00:00
|
|
|
return genpd_finish_resume(dev, pm_generic_thaw_noirq);
|
2017-06-22 07:18:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2017-10-06 07:02:06 +00:00
|
|
|
* genpd_poweroff_noirq - Completion of hibernation of device in an
|
2017-06-22 07:18:33 +00:00
|
|
|
* I/O PM domain.
|
|
|
|
* @dev: Device to poweroff.
|
|
|
|
*
|
|
|
|
* Stop the device and remove power from the domain if all devices in it have
|
|
|
|
* been stopped.
|
|
|
|
*/
|
2017-10-06 07:02:06 +00:00
|
|
|
static int genpd_poweroff_noirq(struct device *dev)
|
2017-06-22 07:18:33 +00:00
|
|
|
{
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
2022-11-02 14:21:02 +00:00
|
|
|
return genpd_finish_suspend(dev,
|
|
|
|
pm_generic_poweroff_noirq,
|
|
|
|
pm_generic_restore_noirq);
|
2011-07-01 20:13:19 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2017-10-06 07:02:06 +00:00
|
|
|
* genpd_restore_noirq - Start of restore of device in an I/O PM domain.
|
2011-07-01 20:13:19 +00:00
|
|
|
* @dev: Device to resume.
|
|
|
|
*
|
2012-01-29 19:39:02 +00:00
|
|
|
* Make sure the domain will be in the same power state as before the
|
|
|
|
* hibernation the system is resuming from and start the device if necessary.
|
2011-07-01 20:13:19 +00:00
|
|
|
*/
|
2017-10-06 07:02:06 +00:00
|
|
|
static int genpd_restore_noirq(struct device *dev)
|
2011-07-01 20:13:19 +00:00
|
|
|
{
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
2022-11-02 14:21:03 +00:00
|
|
|
return genpd_finish_resume(dev, pm_generic_restore_noirq);
|
2011-07-01 20:13:19 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2017-10-06 07:02:06 +00:00
|
|
|
* genpd_complete - Complete power transition of a device in a power domain.
|
2011-07-01 20:13:19 +00:00
|
|
|
* @dev: Device to complete the transition of.
|
|
|
|
*
|
|
|
|
* Complete a power transition of a device (during a system-wide power
|
|
|
|
* transition) under the assumption that its pm_domain field points to the
|
|
|
|
* domain member of an object of type struct generic_pm_domain representing
|
|
|
|
* a power domain consisting of I/O devices.
|
|
|
|
*/
|
2017-10-06 07:02:06 +00:00
|
|
|
static void genpd_complete(struct device *dev)
|
2011-07-01 20:13:19 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
|
|
|
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
|
|
|
genpd = dev_to_genpd(dev);
|
|
|
|
if (IS_ERR(genpd))
|
|
|
|
return;
|
|
|
|
|
PM / Domains: Allow runtime PM during system PM phases
In cases when a PM domain isn't powered off when genpd's ->prepare()
callback is invoked, genpd runtime resumes and disables runtime PM for the
device. This behaviour was needed when genpd managed intermediate states
during the power off sequence, as to maintain proper low power states of
devices during system PM suspend/resume.
Commit ba2bbfbf6307 (PM / Domains: Remove intermediate states from the
power off sequence), enables genpd to improve its behaviour in that
respect.
The PM core disables runtime PM at __device_suspend_late() before it calls
a system PM "late" callback for a device. When resuming a device, after a
corresponding "early" callback has been invoked, the PM core re-enables
runtime PM.
By changing genpd to allow runtime PM according to the same system PM
phases as the PM core, devices can be runtime resumed by their
corresponding subsystem/driver when really needed.
In this way, genpd no longer need to runtime resume the device from its
->prepare() callback. In most cases that avoids unnecessary and energy-
wasting operations of runtime resuming devices that have nothing to do,
only to runtime suspend them shortly after.
Although, because of changing this behaviour in genpd and due to that
genpd powers on the PM domain unconditionally in the system PM resume
"noirq" phase, it could potentially cause a PM domain to stay powered
on even if it's unused after the system has resumed. To avoid this,
schedule a power off work when genpd's system PM ->complete() callback
has been invoked for the last device in the PM domain.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Reviewed-by: Kevin Hilman <khilman@baylibre.com>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2016-05-30 09:33:13 +00:00
|
|
|
pm_generic_complete(dev);
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
2011-07-01 20:13:19 +00:00
|
|
|
|
PM / Domains: Allow genpd to power on during system PM phases
If a PM domain is powered off when the first device starts its system PM
prepare phase, genpd prevents any further attempts to power on the PM
domain during the following system PM phases. Not until the system PM
complete phase is finalized for all devices in the PM domain, genpd again
allows it to be powered on.
This behaviour needs to be changed, as a subsystem/driver for a device in
the same PM domain may still need to be able to serve requests in some of
the system PM phases. Accordingly, it may need to runtime resume its
device and thus also request the corresponding PM domain to be powered on.
To deal with these scenarios, let's make the device operational in the
system PM prepare phase by runtime resuming it, no matter if the PM domain
is powered on or off. Changing this also enables us to remove genpd's
suspend_power_off flag, as it's being used to track this condition.
Additionally, we must allow the PM domain to be powered on via runtime PM
during the system PM phases.
This change also requires a fix in the AMD ACP (Audio CoProcessor) drm
driver. It registers a genpd to model the ACP as a PM domain, but
unfortunately it's also abuses genpd's "internal" suspend_power_off flag
to deal with a corner case at system PM resume.
More precisely, the so called SMU block powers on the ACP at system PM
resume, unconditionally if it's being used or not. This may lead to that
genpd's internal status of the power state, may not correctly reflect the
power state of the HW after a system PM resume.
Because of changing the behaviour of genpd, by runtime resuming devices in
the prepare phase, the AMD ACP drm driver no longer have to deal with this
corner case. So let's just drop the related code in this driver.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Reviewed-by: Kevin Hilman <khilman@baylibre.com>
Acked-by: Maruthi Bayyavarapu <maruthi.bayyavarapu@amd.com>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2016-05-30 09:43:07 +00:00
|
|
|
genpd->prepared_count--;
|
PM / Domains: Allow runtime PM during system PM phases
In cases when a PM domain isn't powered off when genpd's ->prepare()
callback is invoked, genpd runtime resumes and disables runtime PM for the
device. This behaviour was needed when genpd managed intermediate states
during the power off sequence, as to maintain proper low power states of
devices during system PM suspend/resume.
Commit ba2bbfbf6307 (PM / Domains: Remove intermediate states from the
power off sequence), enables genpd to improve its behaviour in that
respect.
The PM core disables runtime PM at __device_suspend_late() before it calls
a system PM "late" callback for a device. When resuming a device, after a
corresponding "early" callback has been invoked, the PM core re-enables
runtime PM.
By changing genpd to allow runtime PM according to the same system PM
phases as the PM core, devices can be runtime resumed by their
corresponding subsystem/driver when really needed.
In this way, genpd no longer need to runtime resume the device from its
->prepare() callback. In most cases that avoids unnecessary and energy-
wasting operations of runtime resuming devices that have nothing to do,
only to runtime suspend them shortly after.
Although, because of changing this behaviour in genpd and due to that
genpd powers on the PM domain unconditionally in the system PM resume
"noirq" phase, it could potentially cause a PM domain to stay powered
on even if it's unused after the system has resumed. To avoid this,
schedule a power off work when genpd's system PM ->complete() callback
has been invoked for the last device in the PM domain.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Reviewed-by: Kevin Hilman <khilman@baylibre.com>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2016-05-30 09:33:13 +00:00
|
|
|
if (!genpd->prepared_count)
|
|
|
|
genpd_queue_power_off_work(genpd);
|
2011-07-01 20:13:19 +00:00
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2011-07-01 20:13:19 +00:00
|
|
|
}
|
|
|
|
|
2020-11-03 15:06:25 +00:00
|
|
|
static void genpd_switch_state(struct device *dev, bool suspend)
|
2012-08-05 23:39:57 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
2020-11-03 15:06:26 +00:00
|
|
|
bool use_lock;
|
2012-08-05 23:39:57 +00:00
|
|
|
|
2019-10-16 14:16:49 +00:00
|
|
|
genpd = dev_to_genpd_safe(dev);
|
|
|
|
if (!genpd)
|
2012-08-05 23:39:57 +00:00
|
|
|
return;
|
|
|
|
|
2020-11-03 15:06:26 +00:00
|
|
|
use_lock = genpd_is_irq_safe(genpd);
|
|
|
|
|
|
|
|
if (use_lock)
|
|
|
|
genpd_lock(genpd);
|
|
|
|
|
2012-08-05 23:39:57 +00:00
|
|
|
if (suspend) {
|
|
|
|
genpd->suspended_count++;
|
2020-11-03 15:06:26 +00:00
|
|
|
genpd_sync_power_off(genpd, use_lock, 0);
|
2012-08-05 23:39:57 +00:00
|
|
|
} else {
|
2020-11-03 15:06:26 +00:00
|
|
|
genpd_sync_power_on(genpd, use_lock, 0);
|
2012-08-05 23:39:57 +00:00
|
|
|
genpd->suspended_count--;
|
|
|
|
}
|
2020-11-03 15:06:26 +00:00
|
|
|
|
|
|
|
if (use_lock)
|
|
|
|
genpd_unlock(genpd);
|
2012-08-05 23:39:57 +00:00
|
|
|
}
|
2014-09-03 10:52:24 +00:00
|
|
|
|
2020-11-03 15:06:25 +00:00
|
|
|
/**
|
|
|
|
* dev_pm_genpd_suspend - Synchronously try to suspend the genpd for @dev
|
|
|
|
* @dev: The device that is attached to the genpd, that can be suspended.
|
|
|
|
*
|
|
|
|
* This routine should typically be called for a device that needs to be
|
2020-11-03 15:06:26 +00:00
|
|
|
* suspended during the syscore suspend phase. It may also be called during
|
|
|
|
* suspend-to-idle to suspend a corresponding CPU device that is attached to a
|
|
|
|
* genpd.
|
2020-11-03 15:06:25 +00:00
|
|
|
*/
|
|
|
|
void dev_pm_genpd_suspend(struct device *dev)
|
2014-09-03 10:52:24 +00:00
|
|
|
{
|
2020-11-03 15:06:25 +00:00
|
|
|
genpd_switch_state(dev, true);
|
2014-09-03 10:52:24 +00:00
|
|
|
}
|
2020-11-03 15:06:25 +00:00
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_suspend);
|
2014-09-03 10:52:24 +00:00
|
|
|
|
2020-11-03 15:06:25 +00:00
|
|
|
/**
|
|
|
|
* dev_pm_genpd_resume - Synchronously try to resume the genpd for @dev
|
|
|
|
* @dev: The device that is attached to the genpd, which needs to be resumed.
|
|
|
|
*
|
|
|
|
* This routine should typically be called for a device that needs to be resumed
|
2020-11-03 15:06:26 +00:00
|
|
|
* during the syscore resume phase. It may also be called during suspend-to-idle
|
|
|
|
* to resume a corresponding CPU device that is attached to a genpd.
|
2020-11-03 15:06:25 +00:00
|
|
|
*/
|
|
|
|
void dev_pm_genpd_resume(struct device *dev)
|
2014-09-03 10:52:24 +00:00
|
|
|
{
|
2020-11-03 15:06:25 +00:00
|
|
|
genpd_switch_state(dev, false);
|
2014-09-03 10:52:24 +00:00
|
|
|
}
|
2020-11-03 15:06:25 +00:00
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_resume);
|
2012-08-05 23:39:57 +00:00
|
|
|
|
2014-11-27 21:38:05 +00:00
|
|
|
#else /* !CONFIG_PM_SLEEP */
|
2011-07-01 20:13:19 +00:00
|
|
|
|
2017-10-06 07:02:06 +00:00
|
|
|
#define genpd_prepare NULL
|
|
|
|
#define genpd_suspend_noirq NULL
|
|
|
|
#define genpd_resume_noirq NULL
|
|
|
|
#define genpd_freeze_noirq NULL
|
|
|
|
#define genpd_thaw_noirq NULL
|
|
|
|
#define genpd_poweroff_noirq NULL
|
|
|
|
#define genpd_restore_noirq NULL
|
|
|
|
#define genpd_complete NULL
|
2011-07-01 20:13:19 +00:00
|
|
|
|
|
|
|
#endif /* CONFIG_PM_SLEEP */
|
|
|
|
|
2022-05-11 14:56:56 +00:00
|
|
|
static struct generic_pm_domain_data *genpd_alloc_dev_data(struct device *dev,
|
|
|
|
bool has_governor)
|
2012-07-05 20:12:32 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain_data *gpd_data;
|
2022-05-11 14:56:56 +00:00
|
|
|
struct gpd_timing_data *td;
|
2015-01-27 20:13:43 +00:00
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = dev_pm_get_subsys_data(dev);
|
|
|
|
if (ret)
|
|
|
|
return ERR_PTR(ret);
|
2012-07-05 20:12:32 +00:00
|
|
|
|
|
|
|
gpd_data = kzalloc(sizeof(*gpd_data), GFP_KERNEL);
|
2015-01-27 20:13:43 +00:00
|
|
|
if (!gpd_data) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto err_put;
|
|
|
|
}
|
2012-07-05 20:12:32 +00:00
|
|
|
|
2015-01-27 20:13:44 +00:00
|
|
|
gpd_data->base.dev = dev;
|
|
|
|
gpd_data->nb.notifier_call = genpd_dev_pm_qos_notifier;
|
|
|
|
|
2022-05-11 14:56:56 +00:00
|
|
|
/* Allocate data used by a governor. */
|
|
|
|
if (has_governor) {
|
|
|
|
td = kzalloc(sizeof(*td), GFP_KERNEL);
|
|
|
|
if (!td) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto err_free;
|
|
|
|
}
|
2015-01-27 20:13:44 +00:00
|
|
|
|
2022-05-11 14:56:56 +00:00
|
|
|
td->constraint_changed = true;
|
|
|
|
td->effective_constraint_ns = PM_QOS_RESUME_LATENCY_NO_CONSTRAINT_NS;
|
2022-05-11 14:56:57 +00:00
|
|
|
td->next_wakeup = KTIME_MAX;
|
2022-05-11 14:56:56 +00:00
|
|
|
gpd_data->td = td;
|
2015-01-27 20:13:44 +00:00
|
|
|
}
|
|
|
|
|
2022-05-11 14:56:56 +00:00
|
|
|
spin_lock_irq(&dev->power.lock);
|
|
|
|
|
|
|
|
if (dev->power.subsys_data->domain_data)
|
|
|
|
ret = -EINVAL;
|
|
|
|
else
|
|
|
|
dev->power.subsys_data->domain_data = &gpd_data->base;
|
2015-01-27 20:13:44 +00:00
|
|
|
|
|
|
|
spin_unlock_irq(&dev->power.lock);
|
|
|
|
|
2022-05-11 14:56:56 +00:00
|
|
|
if (ret)
|
|
|
|
goto err_free;
|
|
|
|
|
2012-07-05 20:12:32 +00:00
|
|
|
return gpd_data;
|
2015-01-27 20:13:43 +00:00
|
|
|
|
2015-01-27 20:13:44 +00:00
|
|
|
err_free:
|
2022-05-11 14:56:56 +00:00
|
|
|
kfree(gpd_data->td);
|
2015-01-27 20:13:44 +00:00
|
|
|
kfree(gpd_data);
|
2015-01-27 20:13:43 +00:00
|
|
|
err_put:
|
|
|
|
dev_pm_put_subsys_data(dev);
|
|
|
|
return ERR_PTR(ret);
|
2012-07-05 20:12:32 +00:00
|
|
|
}
|
|
|
|
|
2015-01-27 20:13:38 +00:00
|
|
|
static void genpd_free_dev_data(struct device *dev,
|
|
|
|
struct generic_pm_domain_data *gpd_data)
|
2012-07-05 20:12:32 +00:00
|
|
|
{
|
2015-01-27 20:13:44 +00:00
|
|
|
spin_lock_irq(&dev->power.lock);
|
|
|
|
|
|
|
|
dev->power.subsys_data->domain_data = NULL;
|
|
|
|
|
|
|
|
spin_unlock_irq(&dev->power.lock);
|
|
|
|
|
2022-05-11 14:56:56 +00:00
|
|
|
kfree(gpd_data->td);
|
2012-07-05 20:12:32 +00:00
|
|
|
kfree(gpd_data);
|
2015-01-27 20:13:43 +00:00
|
|
|
dev_pm_put_subsys_data(dev);
|
2012-07-05 20:12:32 +00:00
|
|
|
}
|
|
|
|
|
2019-04-25 09:04:12 +00:00
|
|
|
static void genpd_update_cpumask(struct generic_pm_domain *genpd,
|
|
|
|
int cpu, bool set, unsigned int depth)
|
2019-03-27 14:35:46 +00:00
|
|
|
{
|
|
|
|
struct gpd_link *link;
|
|
|
|
|
|
|
|
if (!genpd_is_cpu_domain(genpd))
|
|
|
|
return;
|
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry(link, &genpd->child_links, child_node) {
|
|
|
|
struct generic_pm_domain *parent = link->parent;
|
2019-03-27 14:35:46 +00:00
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
genpd_lock_nested(parent, depth + 1);
|
|
|
|
genpd_update_cpumask(parent, cpu, set, depth + 1);
|
|
|
|
genpd_unlock(parent);
|
2019-03-27 14:35:46 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (set)
|
|
|
|
cpumask_set_cpu(cpu, genpd->cpus);
|
|
|
|
else
|
|
|
|
cpumask_clear_cpu(cpu, genpd->cpus);
|
|
|
|
}
|
|
|
|
|
2019-04-25 09:04:12 +00:00
|
|
|
static void genpd_set_cpumask(struct generic_pm_domain *genpd, int cpu)
|
|
|
|
{
|
|
|
|
if (cpu >= 0)
|
|
|
|
genpd_update_cpumask(genpd, cpu, true, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_clear_cpumask(struct generic_pm_domain *genpd, int cpu)
|
|
|
|
{
|
|
|
|
if (cpu >= 0)
|
|
|
|
genpd_update_cpumask(genpd, cpu, false, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int genpd_get_cpu(struct generic_pm_domain *genpd, struct device *dev)
|
2019-03-27 14:35:46 +00:00
|
|
|
{
|
|
|
|
int cpu;
|
|
|
|
|
|
|
|
if (!genpd_is_cpu_domain(genpd))
|
2019-04-25 09:04:12 +00:00
|
|
|
return -1;
|
2019-03-27 14:35:46 +00:00
|
|
|
|
|
|
|
for_each_possible_cpu(cpu) {
|
2019-04-25 09:04:12 +00:00
|
|
|
if (get_cpu_device(cpu) == dev)
|
|
|
|
return cpu;
|
2019-03-27 14:35:46 +00:00
|
|
|
}
|
|
|
|
|
2019-04-25 09:04:12 +00:00
|
|
|
return -1;
|
2019-03-27 14:35:46 +00:00
|
|
|
}
|
|
|
|
|
2019-04-25 09:04:13 +00:00
|
|
|
static int genpd_add_device(struct generic_pm_domain *genpd, struct device *dev,
|
|
|
|
struct device *base_dev)
|
2011-07-01 20:12:45 +00:00
|
|
|
{
|
2022-05-11 14:57:02 +00:00
|
|
|
struct genpd_governor_data *gd = genpd->gd;
|
2015-01-27 20:13:42 +00:00
|
|
|
struct generic_pm_domain_data *gpd_data;
|
2019-04-25 09:04:13 +00:00
|
|
|
int ret;
|
2011-07-01 20:12:45 +00:00
|
|
|
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
2022-05-11 14:57:02 +00:00
|
|
|
gpd_data = genpd_alloc_dev_data(dev, gd);
|
2015-01-27 20:13:43 +00:00
|
|
|
if (IS_ERR(gpd_data))
|
|
|
|
return PTR_ERR(gpd_data);
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
|
2019-04-25 09:04:13 +00:00
|
|
|
gpd_data->cpu = genpd_get_cpu(genpd, base_dev);
|
2019-04-25 09:04:12 +00:00
|
|
|
|
PM: domains: Allow devices attached to genpd to be managed by HW
Some power-domains may be capable of relying on the HW to control the power
for a device that's hooked up to it. Typically, for these kinds of
configurations the consumer driver should be able to change the behavior of
power domain at runtime, control the power domain in SW mode for certain
configurations and handover the control to HW mode for other usecases.
To allow a consumer driver to change the behaviour of the PM domain for its
device, let's provide a new function, dev_pm_genpd_set_hwmode(). Moreover,
let's add a corresponding optional genpd callback, ->set_hwmode_dev(),
which the genpd provider should implement if it can support switching
between HW controlled mode and SW controlled mode. Similarly, add the
dev_pm_genpd_get_hwmode() to allow consumers to read the current mode and
its corresponding optional genpd callback, ->get_hwmode_dev(), which the
genpd provider can also implement to synchronize the initial HW mode
state in genpd_add_device() by reading back the mode from the hardware.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Abel Vesa <abel.vesa@linaro.org>
Signed-off-by: Jagadeesh Kona <quic_jkona@quicinc.com>
Reviewed-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
Reviewed-by: Dhruva Gole <d-gole@ti.com>
Reviewed-by: Taniya Das <quic_tdas@quicinc.com>
Link: https://lore.kernel.org/r/20240624044809.17751-2-quic_jkona@quicinc.com
2024-06-24 04:48:05 +00:00
|
|
|
gpd_data->hw_mode = genpd->get_hwmode_dev ? genpd->get_hwmode_dev(genpd, dev) : false;
|
|
|
|
|
2015-01-27 20:13:45 +00:00
|
|
|
ret = genpd->attach_dev ? genpd->attach_dev(genpd, dev) : 0;
|
|
|
|
if (ret)
|
|
|
|
goto out;
|
2014-09-25 16:28:28 +00:00
|
|
|
|
2019-03-12 06:51:28 +00:00
|
|
|
genpd_lock(genpd);
|
|
|
|
|
2019-04-25 09:04:13 +00:00
|
|
|
genpd_set_cpumask(genpd, gpd_data->cpu);
|
2017-07-14 10:51:48 +00:00
|
|
|
|
2015-01-27 20:13:40 +00:00
|
|
|
genpd->device_count++;
|
2022-05-11 14:57:02 +00:00
|
|
|
if (gd)
|
|
|
|
gd->max_off_time_changed = true;
|
2015-01-27 20:13:40 +00:00
|
|
|
|
2012-07-05 20:12:32 +00:00
|
|
|
list_add_tail(&gpd_data->base.list_node, &genpd->dev_list);
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2024-05-27 14:25:52 +00:00
|
|
|
dev_pm_domain_set(dev, &genpd->domain);
|
2019-03-12 06:51:28 +00:00
|
|
|
out:
|
2015-01-27 20:13:42 +00:00
|
|
|
if (ret)
|
|
|
|
genpd_free_dev_data(dev, gpd_data);
|
|
|
|
else
|
2019-07-04 07:36:17 +00:00
|
|
|
dev_pm_qos_add_notifier(dev, &gpd_data->nb,
|
|
|
|
DEV_PM_QOS_RESUME_LATENCY);
|
2012-07-05 20:12:32 +00:00
|
|
|
|
2011-07-01 20:12:45 +00:00
|
|
|
return ret;
|
|
|
|
}
|
2016-09-12 11:01:11 +00:00
|
|
|
|
|
|
|
/**
|
2018-05-29 10:04:14 +00:00
|
|
|
* pm_genpd_add_device - Add a device to an I/O PM domain.
|
2016-09-12 11:01:11 +00:00
|
|
|
* @genpd: PM domain to add the device to.
|
|
|
|
* @dev: Device to be added.
|
|
|
|
*/
|
2018-05-29 10:04:14 +00:00
|
|
|
int pm_genpd_add_device(struct generic_pm_domain *genpd, struct device *dev)
|
2016-09-12 11:01:11 +00:00
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2023-05-30 09:55:36 +00:00
|
|
|
if (!genpd || !dev)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2016-09-12 11:01:11 +00:00
|
|
|
mutex_lock(&gpd_list_lock);
|
2019-04-25 09:04:13 +00:00
|
|
|
ret = genpd_add_device(genpd, dev, dev);
|
2016-09-12 11:01:11 +00:00
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
2018-05-29 10:04:14 +00:00
|
|
|
EXPORT_SYMBOL_GPL(pm_genpd_add_device);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
2016-09-21 13:38:50 +00:00
|
|
|
static int genpd_remove_device(struct generic_pm_domain *genpd,
|
|
|
|
struct device *dev)
|
2011-07-01 20:12:45 +00:00
|
|
|
{
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
struct generic_pm_domain_data *gpd_data;
|
2011-08-25 13:34:12 +00:00
|
|
|
struct pm_domain_data *pdd;
|
2019-04-25 09:04:13 +00:00
|
|
|
int ret = 0;
|
2011-07-01 20:12:45 +00:00
|
|
|
|
|
|
|
dev_dbg(dev, "%s()\n", __func__);
|
|
|
|
|
2015-01-27 20:13:42 +00:00
|
|
|
pdd = dev->power.subsys_data->domain_data;
|
|
|
|
gpd_data = to_gpd_data(pdd);
|
2019-07-04 07:36:17 +00:00
|
|
|
dev_pm_qos_remove_notifier(dev, &gpd_data->nb,
|
|
|
|
DEV_PM_QOS_RESUME_LATENCY);
|
2015-01-27 20:13:42 +00:00
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
2011-07-01 20:13:19 +00:00
|
|
|
if (genpd->prepared_count > 0) {
|
|
|
|
ret = -EAGAIN;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
genpd->device_count--;
|
2022-05-11 14:57:02 +00:00
|
|
|
if (genpd->gd)
|
|
|
|
genpd->gd->max_off_time_changed = true;
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
|
2019-04-25 09:04:13 +00:00
|
|
|
genpd_clear_cpumask(genpd, gpd_data->cpu);
|
2017-07-14 10:51:48 +00:00
|
|
|
|
2012-05-01 19:33:53 +00:00
|
|
|
list_del_init(&pdd->list_node);
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
|
2024-05-27 14:25:52 +00:00
|
|
|
dev_pm_domain_set(dev, NULL);
|
|
|
|
|
2019-03-12 06:51:28 +00:00
|
|
|
if (genpd->detach_dev)
|
|
|
|
genpd->detach_dev(genpd, dev);
|
|
|
|
|
2015-01-27 20:13:39 +00:00
|
|
|
genpd_free_dev_data(dev, gpd_data);
|
2012-07-05 20:12:32 +00:00
|
|
|
|
PM / Domains: Cache device stop and domain power off governor results, v3
The results of the default device stop and domain power off governor
functions for generic PM domains, default_stop_ok() and
default_power_down_ok(), depend only on the timing data of devices,
which are static, and on their PM QoS constraints. Thus, in theory,
these functions only need to carry out their computations, which may
be time consuming in general, when it is known that the PM QoS
constraint of at least one of the devices in question has changed.
Use the PM QoS notifiers of devices to implement that. First,
introduce new fields, constraint_changed and max_off_time_changed,
into struct gpd_timing_data and struct generic_pm_domain,
respectively, and register a PM QoS notifier function when adding
a device into a domain that will set those fields to 'true' whenever
the device's PM QoS constraint is modified. Second, make
default_stop_ok() and default_power_down_ok() use those fields to
decide whether or not to carry out their computations from scratch.
The device and PM domain hierarchies are taken into account in that
and the expense is that the changes of PM QoS constraints of
suspended devices will not be taken into account immediately, which
isn't guaranteed anyway in general.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
2012-05-01 19:34:07 +00:00
|
|
|
return 0;
|
2011-07-01 20:12:45 +00:00
|
|
|
|
2011-07-01 20:13:19 +00:00
|
|
|
out:
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2019-07-04 07:36:17 +00:00
|
|
|
dev_pm_qos_add_notifier(dev, &gpd_data->nb, DEV_PM_QOS_RESUME_LATENCY);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
2016-09-21 13:38:50 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* pm_genpd_remove_device - Remove a device from an I/O PM domain.
|
|
|
|
* @dev: Device to be removed.
|
|
|
|
*/
|
2018-05-29 10:04:15 +00:00
|
|
|
int pm_genpd_remove_device(struct device *dev)
|
2016-09-21 13:38:50 +00:00
|
|
|
{
|
2019-08-29 14:48:05 +00:00
|
|
|
struct generic_pm_domain *genpd = dev_to_genpd_safe(dev);
|
2018-05-29 10:04:15 +00:00
|
|
|
|
|
|
|
if (!genpd)
|
2016-09-21 13:38:50 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
return genpd_remove_device(genpd, dev);
|
|
|
|
}
|
2015-11-17 19:42:00 +00:00
|
|
|
EXPORT_SYMBOL_GPL(pm_genpd_remove_device);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
/**
|
|
|
|
* dev_pm_genpd_add_notifier - Add a genpd power on/off notifier for @dev
|
|
|
|
*
|
|
|
|
* @dev: Device that should be associated with the notifier
|
|
|
|
* @nb: The notifier block to register
|
|
|
|
*
|
|
|
|
* Users may call this function to add a genpd power on/off notifier for an
|
|
|
|
* attached @dev. Only one notifier per device is allowed. The notifier is
|
|
|
|
* sent when genpd is powering on/off the PM domain.
|
|
|
|
*
|
|
|
|
* It is assumed that the user guarantee that the genpd wouldn't be detached
|
|
|
|
* while this routine is getting called.
|
|
|
|
*
|
|
|
|
* Returns 0 on success and negative error values on failures.
|
|
|
|
*/
|
|
|
|
int dev_pm_genpd_add_notifier(struct device *dev, struct notifier_block *nb)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
|
|
|
struct generic_pm_domain_data *gpd_data;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
genpd = dev_to_genpd_safe(dev);
|
|
|
|
if (!genpd)
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
if (WARN_ON(!dev->power.subsys_data ||
|
|
|
|
!dev->power.subsys_data->domain_data))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
gpd_data = to_gpd_data(dev->power.subsys_data->domain_data);
|
|
|
|
if (gpd_data->power_nb)
|
|
|
|
return -EEXIST;
|
|
|
|
|
|
|
|
genpd_lock(genpd);
|
|
|
|
ret = raw_notifier_chain_register(&genpd->power_notifiers, nb);
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
|
|
|
|
if (ret) {
|
|
|
|
dev_warn(dev, "failed to add notifier for PM domain %s\n",
|
2024-10-30 12:55:10 +00:00
|
|
|
dev_name(&genpd->dev));
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
gpd_data->power_nb = nb;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_add_notifier);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* dev_pm_genpd_remove_notifier - Remove a genpd power on/off notifier for @dev
|
|
|
|
*
|
|
|
|
* @dev: Device that is associated with the notifier
|
|
|
|
*
|
|
|
|
* Users may call this function to remove a genpd power on/off notifier for an
|
|
|
|
* attached @dev.
|
|
|
|
*
|
|
|
|
* It is assumed that the user guarantee that the genpd wouldn't be detached
|
|
|
|
* while this routine is getting called.
|
|
|
|
*
|
|
|
|
* Returns 0 on success and negative error values on failures.
|
|
|
|
*/
|
|
|
|
int dev_pm_genpd_remove_notifier(struct device *dev)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
|
|
|
struct generic_pm_domain_data *gpd_data;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
genpd = dev_to_genpd_safe(dev);
|
|
|
|
if (!genpd)
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
if (WARN_ON(!dev->power.subsys_data ||
|
|
|
|
!dev->power.subsys_data->domain_data))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
gpd_data = to_gpd_data(dev->power.subsys_data->domain_data);
|
|
|
|
if (!gpd_data->power_nb)
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
genpd_lock(genpd);
|
|
|
|
ret = raw_notifier_chain_unregister(&genpd->power_notifiers,
|
|
|
|
gpd_data->power_nb);
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
|
|
|
|
if (ret) {
|
|
|
|
dev_warn(dev, "failed to remove notifier for PM domain %s\n",
|
2024-10-30 12:55:10 +00:00
|
|
|
dev_name(&genpd->dev));
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
gpd_data->power_nb = NULL;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_genpd_remove_notifier);
|
|
|
|
|
2016-09-12 11:01:11 +00:00
|
|
|
static int genpd_add_subdomain(struct generic_pm_domain *genpd,
|
|
|
|
struct generic_pm_domain *subdomain)
|
2011-07-01 20:12:45 +00:00
|
|
|
{
|
2015-10-28 21:19:50 +00:00
|
|
|
struct gpd_link *link, *itr;
|
2011-07-01 20:12:45 +00:00
|
|
|
int ret = 0;
|
|
|
|
|
2012-08-06 23:08:37 +00:00
|
|
|
if (IS_ERR_OR_NULL(genpd) || IS_ERR_OR_NULL(subdomain)
|
|
|
|
|| genpd == subdomain)
|
2011-07-01 20:12:45 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
2016-10-14 17:47:55 +00:00
|
|
|
/*
|
|
|
|
* If the domain can be powered on/off in an IRQ safe
|
|
|
|
* context, ensure that the subdomain can also be
|
|
|
|
* powered on/off in that context.
|
|
|
|
*/
|
|
|
|
if (!genpd_is_irq_safe(genpd) && genpd_is_irq_safe(subdomain)) {
|
2016-11-10 12:52:15 +00:00
|
|
|
WARN(1, "Parent %s of subdomain %s must be IRQ safe\n",
|
2024-10-30 12:55:10 +00:00
|
|
|
dev_name(&genpd->dev), subdomain->name);
|
2016-10-14 17:47:55 +00:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2015-10-28 21:19:50 +00:00
|
|
|
link = kzalloc(sizeof(*link), GFP_KERNEL);
|
|
|
|
if (!link)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(subdomain);
|
|
|
|
genpd_lock_nested(genpd, SINGLE_DEPTH_NESTING);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
2017-03-20 10:19:20 +00:00
|
|
|
if (!genpd_status_on(genpd) && genpd_status_on(subdomain)) {
|
2011-07-01 20:12:45 +00:00
|
|
|
ret = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry(itr, &genpd->parent_links, parent_node) {
|
|
|
|
if (itr->child == subdomain && itr->parent == genpd) {
|
2011-07-01 20:12:45 +00:00
|
|
|
ret = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
link->parent = genpd;
|
|
|
|
list_add_tail(&link->parent_node, &genpd->parent_links);
|
|
|
|
link->child = subdomain;
|
|
|
|
list_add_tail(&link->child_node, &subdomain->child_links);
|
2017-03-20 10:19:20 +00:00
|
|
|
if (genpd_status_on(subdomain))
|
2011-08-08 21:43:04 +00:00
|
|
|
genpd_sd_counter_inc(genpd);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
|
|
|
out:
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
|
|
|
genpd_unlock(subdomain);
|
2015-10-28 21:19:50 +00:00
|
|
|
if (ret)
|
|
|
|
kfree(link);
|
2011-07-01 20:12:45 +00:00
|
|
|
return ret;
|
|
|
|
}
|
2016-09-12 11:01:11 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* pm_genpd_add_subdomain - Add a subdomain to an I/O PM domain.
|
2020-07-08 23:32:13 +00:00
|
|
|
* @genpd: Leader PM domain to add the subdomain to.
|
2016-09-12 11:01:11 +00:00
|
|
|
* @subdomain: Subdomain to be added.
|
|
|
|
*/
|
|
|
|
int pm_genpd_add_subdomain(struct generic_pm_domain *genpd,
|
|
|
|
struct generic_pm_domain *subdomain)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
mutex_lock(&gpd_list_lock);
|
|
|
|
ret = genpd_add_subdomain(genpd, subdomain);
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
2015-10-01 19:22:53 +00:00
|
|
|
EXPORT_SYMBOL_GPL(pm_genpd_add_subdomain);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* pm_genpd_remove_subdomain - Remove a subdomain from an I/O PM domain.
|
2020-07-08 23:32:13 +00:00
|
|
|
* @genpd: Leader PM domain to remove the subdomain from.
|
2011-08-08 21:43:40 +00:00
|
|
|
* @subdomain: Subdomain to be removed.
|
2011-07-01 20:12:45 +00:00
|
|
|
*/
|
|
|
|
int pm_genpd_remove_subdomain(struct generic_pm_domain *genpd,
|
2011-08-08 21:43:40 +00:00
|
|
|
struct generic_pm_domain *subdomain)
|
2011-07-01 20:12:45 +00:00
|
|
|
{
|
2017-06-28 14:56:18 +00:00
|
|
|
struct gpd_link *l, *link;
|
2011-07-01 20:12:45 +00:00
|
|
|
int ret = -EINVAL;
|
|
|
|
|
2011-08-08 21:43:40 +00:00
|
|
|
if (IS_ERR_OR_NULL(genpd) || IS_ERR_OR_NULL(subdomain))
|
2011-07-01 20:12:45 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(subdomain);
|
|
|
|
genpd_lock_nested(genpd, SINGLE_DEPTH_NESTING);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
if (!list_empty(&subdomain->parent_links) || subdomain->device_count) {
|
2019-03-04 17:14:38 +00:00
|
|
|
pr_warn("%s: unable to remove subdomain %s\n",
|
2024-10-30 12:55:10 +00:00
|
|
|
dev_name(&genpd->dev), subdomain->name);
|
2015-09-03 08:10:37 +00:00
|
|
|
ret = -EBUSY;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry_safe(link, l, &genpd->parent_links, parent_node) {
|
|
|
|
if (link->child != subdomain)
|
2011-07-01 20:12:45 +00:00
|
|
|
continue;
|
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
list_del(&link->parent_node);
|
|
|
|
list_del(&link->child_node);
|
2011-08-08 21:43:40 +00:00
|
|
|
kfree(link);
|
2017-03-20 10:19:20 +00:00
|
|
|
if (genpd_status_on(subdomain))
|
2011-07-01 20:12:45 +00:00
|
|
|
genpd_sd_counter_dec(genpd);
|
|
|
|
|
|
|
|
ret = 0;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2015-09-03 08:10:37 +00:00
|
|
|
out:
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
|
|
|
genpd_unlock(subdomain);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
2015-10-01 19:22:53 +00:00
|
|
|
EXPORT_SYMBOL_GPL(pm_genpd_remove_subdomain);
|
2011-07-01 20:12:45 +00:00
|
|
|
|
2019-03-27 14:35:45 +00:00
|
|
|
static void genpd_free_default_power_state(struct genpd_power_state *states,
|
|
|
|
unsigned int state_count)
|
|
|
|
{
|
|
|
|
kfree(states);
|
|
|
|
}
|
|
|
|
|
2016-10-14 17:47:49 +00:00
|
|
|
static int genpd_set_default_power_state(struct generic_pm_domain *genpd)
|
|
|
|
{
|
|
|
|
struct genpd_power_state *state;
|
|
|
|
|
|
|
|
state = kzalloc(sizeof(*state), GFP_KERNEL);
|
|
|
|
if (!state)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
genpd->states = state;
|
|
|
|
genpd->state_count = 1;
|
2019-03-27 14:35:45 +00:00
|
|
|
genpd->free_states = genpd_free_default_power_state;
|
2016-10-14 17:47:49 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2022-05-11 14:57:01 +00:00
|
|
|
static int genpd_alloc_data(struct generic_pm_domain *genpd)
|
|
|
|
{
|
2022-05-11 14:57:02 +00:00
|
|
|
struct genpd_governor_data *gd = NULL;
|
2022-05-11 14:57:01 +00:00
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (genpd_is_cpu_domain(genpd) &&
|
|
|
|
!zalloc_cpumask_var(&genpd->cpus, GFP_KERNEL))
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2022-05-11 14:57:02 +00:00
|
|
|
if (genpd->gov) {
|
|
|
|
gd = kzalloc(sizeof(*gd), GFP_KERNEL);
|
|
|
|
if (!gd) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto free;
|
|
|
|
}
|
|
|
|
|
|
|
|
gd->max_off_time_ns = -1;
|
|
|
|
gd->max_off_time_changed = true;
|
|
|
|
gd->next_wakeup = KTIME_MAX;
|
2022-10-18 15:28:35 +00:00
|
|
|
gd->next_hrtimer = KTIME_MAX;
|
2022-05-11 14:57:02 +00:00
|
|
|
}
|
|
|
|
|
2022-05-11 14:57:01 +00:00
|
|
|
/* Use only one "off" state if there were no states declared */
|
|
|
|
if (genpd->state_count == 0) {
|
|
|
|
ret = genpd_set_default_power_state(genpd);
|
|
|
|
if (ret)
|
|
|
|
goto free;
|
|
|
|
}
|
|
|
|
|
2022-05-11 14:57:02 +00:00
|
|
|
genpd->gd = gd;
|
2022-05-11 14:57:01 +00:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
free:
|
|
|
|
if (genpd_is_cpu_domain(genpd))
|
|
|
|
free_cpumask_var(genpd->cpus);
|
2022-05-11 14:57:02 +00:00
|
|
|
kfree(gd);
|
2022-05-11 14:57:01 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void genpd_free_data(struct generic_pm_domain *genpd)
|
|
|
|
{
|
|
|
|
if (genpd_is_cpu_domain(genpd))
|
|
|
|
free_cpumask_var(genpd->cpus);
|
|
|
|
if (genpd->free_states)
|
|
|
|
genpd->free_states(genpd->states, genpd->state_count);
|
2022-05-11 14:57:02 +00:00
|
|
|
kfree(genpd->gd);
|
2022-05-11 14:57:01 +00:00
|
|
|
}
|
|
|
|
|
2016-10-14 17:47:55 +00:00
|
|
|
static void genpd_lock_init(struct generic_pm_domain *genpd)
|
|
|
|
{
|
2024-05-27 14:25:51 +00:00
|
|
|
if (genpd_is_cpu_domain(genpd)) {
|
|
|
|
raw_spin_lock_init(&genpd->raw_slock);
|
|
|
|
genpd->lock_ops = &genpd_raw_spin_ops;
|
|
|
|
} else if (genpd_is_irq_safe(genpd)) {
|
2016-10-14 17:47:55 +00:00
|
|
|
spin_lock_init(&genpd->slock);
|
|
|
|
genpd->lock_ops = &genpd_spin_ops;
|
|
|
|
} else {
|
|
|
|
mutex_init(&genpd->mlock);
|
|
|
|
genpd->lock_ops = &genpd_mtx_ops;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-07-01 20:12:45 +00:00
|
|
|
/**
|
|
|
|
* pm_genpd_init - Initialize a generic I/O PM domain object.
|
|
|
|
* @genpd: PM domain object to initialize.
|
|
|
|
* @gov: PM domain governor to associate with the domain (may be NULL).
|
|
|
|
* @is_off: Initial value of the domain's power_is_off field.
|
2016-06-17 10:27:52 +00:00
|
|
|
*
|
|
|
|
* Returns 0 on successful initialization, else a negative error code.
|
2011-07-01 20:12:45 +00:00
|
|
|
*/
|
2016-06-17 10:27:52 +00:00
|
|
|
int pm_genpd_init(struct generic_pm_domain *genpd,
|
|
|
|
struct dev_power_governor *gov, bool is_off)
|
2011-07-01 20:12:45 +00:00
|
|
|
{
|
2016-10-14 17:47:49 +00:00
|
|
|
int ret;
|
|
|
|
|
2011-07-01 20:12:45 +00:00
|
|
|
if (IS_ERR_OR_NULL(genpd))
|
2016-06-17 10:27:52 +00:00
|
|
|
return -EINVAL;
|
2011-07-01 20:12:45 +00:00
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
INIT_LIST_HEAD(&genpd->parent_links);
|
|
|
|
INIT_LIST_HEAD(&genpd->child_links);
|
2011-07-01 20:12:45 +00:00
|
|
|
INIT_LIST_HEAD(&genpd->dev_list);
|
PM: domains: Add support for PM domain on/off notifiers for genpd
A device may have specific HW constraints that must be obeyed to, before
its corresponding PM domain (genpd) can be powered off - and vice verse at
power on. These constraints can't be managed through the regular runtime PM
based deployment for a device, because the access pattern for it, isn't
always request based. In other words, using the runtime PM callbacks to
deal with the constraints doesn't work for these cases.
For these reasons, let's instead add a PM domain power on/off notification
mechanism to genpd. To add/remove a notifier for a device, the device must
already have been attached to the genpd, which also means that it needs to
be a part of the PM domain topology.
To add/remove a notifier, let's introduce two genpd specific functions:
- dev_pm_genpd_add|remove_notifier()
Note that, to further clarify when genpd power on/off notifiers may be
used, one can compare with the existing CPU_CLUSTER_PM_ENTER|EXIT
notifiers. In the long run, the genpd power on/off notifiers should be able
to replace them, but that requires additional genpd based platform support
for the current users.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Tested-by: Lina Iyer <ilina@codeaurora.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2020-10-13 12:23:39 +00:00
|
|
|
RAW_INIT_NOTIFIER_HEAD(&genpd->power_notifiers);
|
2016-10-14 17:47:55 +00:00
|
|
|
genpd_lock_init(genpd);
|
2011-07-01 20:12:45 +00:00
|
|
|
genpd->gov = gov;
|
|
|
|
INIT_WORK(&genpd->power_off_work, genpd_power_off_work_fn);
|
2011-08-08 21:43:04 +00:00
|
|
|
atomic_set(&genpd->sd_count, 0);
|
2020-09-24 11:04:47 +00:00
|
|
|
genpd->status = is_off ? GENPD_STATE_OFF : GENPD_STATE_ON;
|
2011-07-01 20:13:19 +00:00
|
|
|
genpd->device_count = 0;
|
2016-09-12 11:01:12 +00:00
|
|
|
genpd->provider = NULL;
|
2024-10-30 12:55:10 +00:00
|
|
|
genpd->device_id = -ENXIO;
|
2016-09-12 11:01:12 +00:00
|
|
|
genpd->has_provider = false;
|
2022-04-19 17:29:16 +00:00
|
|
|
genpd->accounting_time = ktime_get_mono_fast_ns();
|
2016-03-31 09:21:26 +00:00
|
|
|
genpd->domain.ops.runtime_suspend = genpd_runtime_suspend;
|
|
|
|
genpd->domain.ops.runtime_resume = genpd_runtime_resume;
|
2017-10-06 07:02:06 +00:00
|
|
|
genpd->domain.ops.prepare = genpd_prepare;
|
|
|
|
genpd->domain.ops.suspend_noirq = genpd_suspend_noirq;
|
|
|
|
genpd->domain.ops.resume_noirq = genpd_resume_noirq;
|
|
|
|
genpd->domain.ops.freeze_noirq = genpd_freeze_noirq;
|
|
|
|
genpd->domain.ops.thaw_noirq = genpd_thaw_noirq;
|
|
|
|
genpd->domain.ops.poweroff_noirq = genpd_poweroff_noirq;
|
|
|
|
genpd->domain.ops.restore_noirq = genpd_restore_noirq;
|
|
|
|
genpd->domain.ops.complete = genpd_complete;
|
2019-10-16 13:16:24 +00:00
|
|
|
genpd->domain.start = genpd_dev_pm_start;
|
2023-09-25 13:17:08 +00:00
|
|
|
genpd->domain.set_performance_state = genpd_dev_pm_set_performance_state;
|
2014-12-01 11:50:21 +00:00
|
|
|
|
|
|
|
if (genpd->flags & GENPD_FLAG_PM_CLK) {
|
|
|
|
genpd->dev_ops.stop = pm_clk_suspend;
|
|
|
|
genpd->dev_ops.start = pm_clk_resume;
|
|
|
|
}
|
|
|
|
|
2022-05-11 14:56:51 +00:00
|
|
|
/* The always-on governor works better with the corresponding flag. */
|
|
|
|
if (gov == &pm_domain_always_on_gov)
|
|
|
|
genpd->flags |= GENPD_FLAG_RPM_ALWAYS_ON;
|
|
|
|
|
2017-03-20 10:19:21 +00:00
|
|
|
/* Always-on domains must be powered on at initialization. */
|
2019-04-30 15:06:11 +00:00
|
|
|
if ((genpd_is_always_on(genpd) || genpd_is_rpm_always_on(genpd)) &&
|
2022-09-29 15:42:14 +00:00
|
|
|
!genpd_status_on(genpd)) {
|
|
|
|
pr_err("always-on PM domain %s is not on\n", genpd->name);
|
2017-03-20 10:19:21 +00:00
|
|
|
return -EINVAL;
|
2022-09-29 15:42:14 +00:00
|
|
|
}
|
2017-03-20 10:19:21 +00:00
|
|
|
|
2022-05-11 14:57:01 +00:00
|
|
|
/* Multiple states but no governor doesn't make sense. */
|
|
|
|
if (!gov && genpd->state_count > 1)
|
2019-03-04 17:14:38 +00:00
|
|
|
pr_warn("%s: no governor for states\n", genpd->name);
|
2022-05-11 14:57:01 +00:00
|
|
|
|
|
|
|
ret = genpd_alloc_data(genpd);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
2016-02-15 10:10:51 +00:00
|
|
|
|
2017-03-17 05:56:19 +00:00
|
|
|
device_initialize(&genpd->dev);
|
2024-10-30 12:55:10 +00:00
|
|
|
|
|
|
|
if (!genpd_is_dev_name_fw(genpd)) {
|
|
|
|
dev_set_name(&genpd->dev, "%s", genpd->name);
|
|
|
|
} else {
|
|
|
|
ret = ida_alloc(&genpd_ida, GFP_KERNEL);
|
|
|
|
if (ret < 0) {
|
|
|
|
put_device(&genpd->dev);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
genpd->device_id = ret;
|
|
|
|
dev_set_name(&genpd->dev, "%s_%u", genpd->name, genpd->device_id);
|
|
|
|
}
|
2017-03-17 05:56:19 +00:00
|
|
|
|
2011-07-13 10:31:52 +00:00
|
|
|
mutex_lock(&gpd_list_lock);
|
|
|
|
list_add(&genpd->gpd_list_node, &gpd_list);
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
2021-06-24 20:18:02 +00:00
|
|
|
genpd_debug_add(genpd);
|
2016-06-17 10:27:52 +00:00
|
|
|
|
|
|
|
return 0;
|
2011-07-13 10:31:52 +00:00
|
|
|
}
|
2015-08-13 06:21:57 +00:00
|
|
|
EXPORT_SYMBOL_GPL(pm_genpd_init);
|
2014-09-19 18:27:36 +00:00
|
|
|
|
2016-09-12 11:01:13 +00:00
|
|
|
static int genpd_remove(struct generic_pm_domain *genpd)
|
|
|
|
{
|
|
|
|
struct gpd_link *l, *link;
|
|
|
|
|
|
|
|
if (IS_ERR_OR_NULL(genpd))
|
|
|
|
return -EINVAL;
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_lock(genpd);
|
2016-09-12 11:01:13 +00:00
|
|
|
|
|
|
|
if (genpd->has_provider) {
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2024-10-30 12:55:10 +00:00
|
|
|
pr_err("Provider present, unable to remove %s\n", dev_name(&genpd->dev));
|
2016-09-12 11:01:13 +00:00
|
|
|
return -EBUSY;
|
|
|
|
}
|
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
if (!list_empty(&genpd->parent_links) || genpd->device_count) {
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2024-10-30 12:55:10 +00:00
|
|
|
pr_err("%s: unable to remove %s\n", __func__, dev_name(&genpd->dev));
|
2016-09-12 11:01:13 +00:00
|
|
|
return -EBUSY;
|
|
|
|
}
|
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry_safe(link, l, &genpd->child_links, child_node) {
|
|
|
|
list_del(&link->parent_node);
|
|
|
|
list_del(&link->child_node);
|
2016-09-12 11:01:13 +00:00
|
|
|
kfree(link);
|
|
|
|
}
|
|
|
|
|
|
|
|
list_del(&genpd->gpd_list_node);
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
2022-02-25 06:48:15 +00:00
|
|
|
genpd_debug_remove(genpd);
|
2016-09-12 11:01:13 +00:00
|
|
|
cancel_work_sync(&genpd->power_off_work);
|
2024-10-30 12:55:10 +00:00
|
|
|
if (genpd->device_id != -ENXIO)
|
|
|
|
ida_free(&genpd_ida, genpd->device_id);
|
2022-05-11 14:57:01 +00:00
|
|
|
genpd_free_data(genpd);
|
2019-03-27 14:35:45 +00:00
|
|
|
|
2024-10-30 12:55:10 +00:00
|
|
|
pr_debug("%s: removed %s\n", __func__, dev_name(&genpd->dev));
|
2016-09-12 11:01:13 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* pm_genpd_remove - Remove a generic I/O PM domain
|
|
|
|
* @genpd: Pointer to PM domain that is to be removed.
|
|
|
|
*
|
|
|
|
* To remove the PM domain, this function:
|
|
|
|
* - Removes the PM domain as a subdomain to any parent domains,
|
|
|
|
* if it was added.
|
|
|
|
* - Removes the PM domain from the list of registered PM domains.
|
|
|
|
*
|
|
|
|
* The PM domain will only be removed, if the associated provider has
|
|
|
|
* been removed, it is not a parent to any other PM domain and has no
|
|
|
|
* devices associated with it.
|
|
|
|
*/
|
|
|
|
int pm_genpd_remove(struct generic_pm_domain *genpd)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
mutex_lock(&gpd_list_lock);
|
|
|
|
ret = genpd_remove(genpd);
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(pm_genpd_remove);
|
|
|
|
|
2014-09-19 18:27:36 +00:00
|
|
|
#ifdef CONFIG_PM_GENERIC_DOMAINS_OF
|
2016-09-12 11:01:09 +00:00
|
|
|
|
2014-09-19 18:27:36 +00:00
|
|
|
/*
|
|
|
|
* Device Tree based PM domain providers.
|
|
|
|
*
|
|
|
|
* The code below implements generic device tree based PM domain providers that
|
|
|
|
* bind device tree nodes with generic PM domains registered in the system.
|
|
|
|
*
|
|
|
|
* Any driver that registers generic PM domains and needs to support binding of
|
|
|
|
* devices to these domains is supposed to register a PM domain provider, which
|
|
|
|
* maps a PM domain specifier retrieved from the device tree to a PM domain.
|
|
|
|
*
|
|
|
|
* Two simple mapping functions have been provided for convenience:
|
2016-09-12 11:01:09 +00:00
|
|
|
* - genpd_xlate_simple() for 1:1 device tree node to PM domain mapping.
|
|
|
|
* - genpd_xlate_onecell() for mapping of multiple PM domains per node by
|
2014-09-19 18:27:36 +00:00
|
|
|
* index.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
* struct of_genpd_provider - PM domain provider registration structure
|
|
|
|
* @link: Entry in global list of PM domain providers
|
|
|
|
* @node: Pointer to device tree node of PM domain provider
|
|
|
|
* @xlate: Provider-specific xlate callback mapping a set of specifier cells
|
|
|
|
* into a PM domain.
|
|
|
|
* @data: context pointer to be passed into @xlate callback
|
|
|
|
*/
|
|
|
|
struct of_genpd_provider {
|
|
|
|
struct list_head link;
|
|
|
|
struct device_node *node;
|
|
|
|
genpd_xlate_t xlate;
|
|
|
|
void *data;
|
|
|
|
};
|
|
|
|
|
|
|
|
/* List of registered PM domain providers. */
|
|
|
|
static LIST_HEAD(of_genpd_providers);
|
|
|
|
/* Mutex to protect the list above. */
|
|
|
|
static DEFINE_MUTEX(of_genpd_mutex);
|
|
|
|
|
|
|
|
/**
|
2016-09-12 11:01:09 +00:00
|
|
|
* genpd_xlate_simple() - Xlate function for direct node-domain mapping
|
2014-09-19 18:27:36 +00:00
|
|
|
* @genpdspec: OF phandle args to map into a PM domain
|
|
|
|
* @data: xlate function private data - pointer to struct generic_pm_domain
|
|
|
|
*
|
|
|
|
* This is a generic xlate function that can be used to model PM domains that
|
|
|
|
* have their own device tree nodes. The private data of xlate function needs
|
|
|
|
* to be a valid pointer to struct generic_pm_domain.
|
|
|
|
*/
|
2016-09-12 11:01:09 +00:00
|
|
|
static struct generic_pm_domain *genpd_xlate_simple(
|
2024-02-08 20:28:21 +00:00
|
|
|
const struct of_phandle_args *genpdspec,
|
2014-09-19 18:27:36 +00:00
|
|
|
void *data)
|
|
|
|
{
|
|
|
|
return data;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2016-09-12 11:01:09 +00:00
|
|
|
* genpd_xlate_onecell() - Xlate function using a single index.
|
2014-09-19 18:27:36 +00:00
|
|
|
* @genpdspec: OF phandle args to map into a PM domain
|
|
|
|
* @data: xlate function private data - pointer to struct genpd_onecell_data
|
|
|
|
*
|
|
|
|
* This is a generic xlate function that can be used to model simple PM domain
|
|
|
|
* controllers that have one device tree node and provide multiple PM domains.
|
|
|
|
* A single cell is used as an index into an array of PM domains specified in
|
|
|
|
* the genpd_onecell_data struct when registering the provider.
|
|
|
|
*/
|
2016-09-12 11:01:09 +00:00
|
|
|
static struct generic_pm_domain *genpd_xlate_onecell(
|
2024-02-08 20:28:21 +00:00
|
|
|
const struct of_phandle_args *genpdspec,
|
2014-09-19 18:27:36 +00:00
|
|
|
void *data)
|
|
|
|
{
|
|
|
|
struct genpd_onecell_data *genpd_data = data;
|
|
|
|
unsigned int idx = genpdspec->args[0];
|
|
|
|
|
|
|
|
if (genpdspec->args_count != 1)
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
|
|
|
if (idx >= genpd_data->num_domains) {
|
|
|
|
pr_err("%s: invalid domain index %u\n", __func__, idx);
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!genpd_data->domains[idx])
|
|
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
|
|
|
|
return genpd_data->domains[idx];
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2016-09-12 11:01:09 +00:00
|
|
|
* genpd_add_provider() - Register a PM domain provider for a node
|
2014-09-19 18:27:36 +00:00
|
|
|
* @np: Device node pointer associated with the PM domain provider.
|
|
|
|
* @xlate: Callback for decoding PM domain from phandle arguments.
|
|
|
|
* @data: Context pointer for @xlate callback.
|
|
|
|
*/
|
2016-09-12 11:01:09 +00:00
|
|
|
static int genpd_add_provider(struct device_node *np, genpd_xlate_t xlate,
|
|
|
|
void *data)
|
2014-09-19 18:27:36 +00:00
|
|
|
{
|
|
|
|
struct of_genpd_provider *cp;
|
|
|
|
|
|
|
|
cp = kzalloc(sizeof(*cp), GFP_KERNEL);
|
|
|
|
if (!cp)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
cp->node = of_node_get(np);
|
|
|
|
cp->data = data;
|
|
|
|
cp->xlate = xlate;
|
2021-02-05 22:26:43 +00:00
|
|
|
fwnode_dev_initialized(&np->fwnode, true);
|
2014-09-19 18:27:36 +00:00
|
|
|
|
|
|
|
mutex_lock(&of_genpd_mutex);
|
|
|
|
list_add(&cp->link, &of_genpd_providers);
|
|
|
|
mutex_unlock(&of_genpd_mutex);
|
2017-07-18 21:42:50 +00:00
|
|
|
pr_debug("Added domain provider from %pOF\n", np);
|
2014-09-19 18:27:36 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
2016-09-12 11:01:09 +00:00
|
|
|
|
2019-10-16 14:16:49 +00:00
|
|
|
static bool genpd_present(const struct generic_pm_domain *genpd)
|
|
|
|
{
|
2021-06-24 20:18:02 +00:00
|
|
|
bool ret = false;
|
2019-10-16 14:16:49 +00:00
|
|
|
const struct generic_pm_domain *gpd;
|
|
|
|
|
2021-06-24 20:18:02 +00:00
|
|
|
mutex_lock(&gpd_list_lock);
|
|
|
|
list_for_each_entry(gpd, &gpd_list, gpd_list_node) {
|
|
|
|
if (gpd == genpd) {
|
|
|
|
ret = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
|
|
|
return ret;
|
2019-10-16 14:16:49 +00:00
|
|
|
}
|
|
|
|
|
2016-09-12 11:01:09 +00:00
|
|
|
/**
|
|
|
|
* of_genpd_add_provider_simple() - Register a simple PM domain provider
|
|
|
|
* @np: Device node pointer associated with the PM domain provider.
|
|
|
|
* @genpd: Pointer to PM domain associated with the PM domain provider.
|
|
|
|
*/
|
|
|
|
int of_genpd_add_provider_simple(struct device_node *np,
|
|
|
|
struct generic_pm_domain *genpd)
|
|
|
|
{
|
2021-06-24 20:18:02 +00:00
|
|
|
int ret;
|
2016-09-12 11:01:10 +00:00
|
|
|
|
|
|
|
if (!np || !genpd)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2018-04-05 10:23:34 +00:00
|
|
|
if (!genpd_present(genpd))
|
2021-06-24 20:18:02 +00:00
|
|
|
return -EINVAL;
|
2018-04-05 10:23:34 +00:00
|
|
|
|
|
|
|
genpd->dev.of_node = np;
|
|
|
|
|
|
|
|
/* Parse genpd OPP table */
|
2023-08-25 11:26:32 +00:00
|
|
|
if (!genpd_is_opp_table_fw(genpd) && genpd->set_performance_state) {
|
2018-04-05 10:23:34 +00:00
|
|
|
ret = dev_pm_opp_of_add_table(&genpd->dev);
|
2022-02-23 08:03:23 +00:00
|
|
|
if (ret)
|
|
|
|
return dev_err_probe(&genpd->dev, ret, "Failed to add OPP table\n");
|
2018-11-02 05:48:08 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Save table for faster processing while setting performance
|
|
|
|
* state.
|
|
|
|
*/
|
|
|
|
genpd->opp_table = dev_pm_opp_get_opp_table(&genpd->dev);
|
opp: Allow dev_pm_opp_get_opp_table() to return -EPROBE_DEFER
The OPP core manages various resources, e.g. clocks or interconnect paths.
These resources are looked up when the OPP table is allocated once
dev_pm_opp_get_opp_table() is called the first time (either directly
or indirectly through one of the many helper functions).
At this point, the resources may not be available yet, i.e. looking them
up will result in -EPROBE_DEFER. Unfortunately, dev_pm_opp_get_opp_table()
is currently unable to propagate this error code since it only returns
the allocated OPP table or NULL.
This means that all consumers of the OPP core are required to make sure
that all necessary resources are available. Usually this happens by
requesting them, checking the result and releasing them immediately after.
For example, we have added "dev_pm_opp_of_find_icc_paths(dev, NULL)" to
several drivers now just to make sure the interconnect providers are
ready before the OPP table is allocated. If this call is missing,
the OPP core will only warn about this and then attempt to continue
without interconnect. This will eventually fail horribly, e.g.:
cpu cpu0: _allocate_opp_table: Error finding interconnect paths: -517
... later ...
of: _read_bw: Mismatch between opp-peak-kBps and paths (1 0)
cpu cpu0: _opp_add_static_v2: opp key field not found
cpu cpu0: _of_add_opp_table_v2: Failed to add OPP, -22
This example happens when trying to use interconnects for a CPU OPP
table together with qcom-cpufreq-nvmem.c. qcom-cpufreq-nvmem calls
dev_pm_opp_set_supported_hw(), which ends up allocating the OPP table
early. To fix the problem with the current approach we would need to add
yet another call to dev_pm_opp_of_find_icc_paths(dev, NULL).
But actually qcom-cpufreq-nvmem.c has nothing to do with interconnects...
This commit attempts to make this more robust by allowing
dev_pm_opp_get_opp_table() to return an error pointer. Fixing all
the usages is trivial because the function is usually used indirectly
through another helper (e.g. dev_pm_opp_set_supported_hw() above).
These other helpers already return an error pointer.
The example above then works correctly because set_supported_hw() will
return -EPROBE_DEFER, and qcom-cpufreq-nvmem.c already propagates that
error. It should also be possible to remove the remaining usages of
"dev_pm_opp_of_find_icc_paths(dev, NULL)" from other drivers as well.
Note that this commit currently only handles -EPROBE_DEFER for the
clock/interconnects within _allocate_opp_table(). Other errors are just
ignored as before. Eventually those should be propagated as well.
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Acked-by: Krzysztof Kozlowski <krzk@kernel.org>
Reviewed-by: Ulf Hansson <ulf.hansson@linaro.org>
[ Viresh: skip checking return value of dev_pm_opp_get_opp_table() for
EPROBE_DEFER in domain.c, fix NULL return value and reorder
code a bit in core.c, and update exynos-asv.c ]
Signed-off-by: Viresh Kumar <viresh.kumar@linaro.org>
2020-07-27 09:30:46 +00:00
|
|
|
WARN_ON(IS_ERR(genpd->opp_table));
|
2016-09-12 11:01:12 +00:00
|
|
|
}
|
|
|
|
|
2018-04-05 10:23:34 +00:00
|
|
|
ret = genpd_add_provider(np, genpd_xlate_simple, genpd);
|
|
|
|
if (ret) {
|
2023-08-25 11:26:32 +00:00
|
|
|
if (!genpd_is_opp_table_fw(genpd) && genpd->set_performance_state) {
|
2018-11-02 05:48:08 +00:00
|
|
|
dev_pm_opp_put_opp_table(genpd->opp_table);
|
2018-04-05 10:23:34 +00:00
|
|
|
dev_pm_opp_of_remove_table(&genpd->dev);
|
2018-11-02 05:48:08 +00:00
|
|
|
}
|
2018-04-05 10:23:34 +00:00
|
|
|
|
2021-06-24 20:18:02 +00:00
|
|
|
return ret;
|
2018-04-05 10:23:34 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
genpd->provider = &np->fwnode;
|
|
|
|
genpd->has_provider = true;
|
|
|
|
|
2021-06-24 20:18:02 +00:00
|
|
|
return 0;
|
2016-09-12 11:01:09 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(of_genpd_add_provider_simple);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* of_genpd_add_provider_onecell() - Register a onecell PM domain provider
|
|
|
|
* @np: Device node pointer associated with the PM domain provider.
|
|
|
|
* @data: Pointer to the data associated with the PM domain provider.
|
|
|
|
*/
|
|
|
|
int of_genpd_add_provider_onecell(struct device_node *np,
|
|
|
|
struct genpd_onecell_data *data)
|
|
|
|
{
|
2018-04-05 10:23:34 +00:00
|
|
|
struct generic_pm_domain *genpd;
|
2016-09-12 11:01:10 +00:00
|
|
|
unsigned int i;
|
2016-09-12 11:01:12 +00:00
|
|
|
int ret = -EINVAL;
|
2016-09-12 11:01:10 +00:00
|
|
|
|
|
|
|
if (!np || !data)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2017-03-29 16:34:50 +00:00
|
|
|
if (!data->xlate)
|
|
|
|
data->xlate = genpd_xlate_onecell;
|
|
|
|
|
2016-09-12 11:01:10 +00:00
|
|
|
for (i = 0; i < data->num_domains; i++) {
|
2018-04-05 10:23:34 +00:00
|
|
|
genpd = data->domains[i];
|
|
|
|
|
|
|
|
if (!genpd)
|
2016-09-15 12:05:23 +00:00
|
|
|
continue;
|
2018-04-05 10:23:34 +00:00
|
|
|
if (!genpd_present(genpd))
|
2016-09-12 11:01:12 +00:00
|
|
|
goto error;
|
|
|
|
|
2018-04-05 10:23:34 +00:00
|
|
|
genpd->dev.of_node = np;
|
|
|
|
|
|
|
|
/* Parse genpd OPP table */
|
2023-08-25 11:26:32 +00:00
|
|
|
if (!genpd_is_opp_table_fw(genpd) && genpd->set_performance_state) {
|
2018-04-05 10:23:34 +00:00
|
|
|
ret = dev_pm_opp_of_add_table_indexed(&genpd->dev, i);
|
|
|
|
if (ret) {
|
2022-02-23 08:03:23 +00:00
|
|
|
dev_err_probe(&genpd->dev, ret,
|
|
|
|
"Failed to add OPP table for index %d\n", i);
|
2018-04-05 10:23:34 +00:00
|
|
|
goto error;
|
|
|
|
}
|
2018-11-02 05:48:08 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Save table for faster processing while setting
|
|
|
|
* performance state.
|
|
|
|
*/
|
2020-11-06 05:07:16 +00:00
|
|
|
genpd->opp_table = dev_pm_opp_get_opp_table(&genpd->dev);
|
opp: Allow dev_pm_opp_get_opp_table() to return -EPROBE_DEFER
The OPP core manages various resources, e.g. clocks or interconnect paths.
These resources are looked up when the OPP table is allocated once
dev_pm_opp_get_opp_table() is called the first time (either directly
or indirectly through one of the many helper functions).
At this point, the resources may not be available yet, i.e. looking them
up will result in -EPROBE_DEFER. Unfortunately, dev_pm_opp_get_opp_table()
is currently unable to propagate this error code since it only returns
the allocated OPP table or NULL.
This means that all consumers of the OPP core are required to make sure
that all necessary resources are available. Usually this happens by
requesting them, checking the result and releasing them immediately after.
For example, we have added "dev_pm_opp_of_find_icc_paths(dev, NULL)" to
several drivers now just to make sure the interconnect providers are
ready before the OPP table is allocated. If this call is missing,
the OPP core will only warn about this and then attempt to continue
without interconnect. This will eventually fail horribly, e.g.:
cpu cpu0: _allocate_opp_table: Error finding interconnect paths: -517
... later ...
of: _read_bw: Mismatch between opp-peak-kBps and paths (1 0)
cpu cpu0: _opp_add_static_v2: opp key field not found
cpu cpu0: _of_add_opp_table_v2: Failed to add OPP, -22
This example happens when trying to use interconnects for a CPU OPP
table together with qcom-cpufreq-nvmem.c. qcom-cpufreq-nvmem calls
dev_pm_opp_set_supported_hw(), which ends up allocating the OPP table
early. To fix the problem with the current approach we would need to add
yet another call to dev_pm_opp_of_find_icc_paths(dev, NULL).
But actually qcom-cpufreq-nvmem.c has nothing to do with interconnects...
This commit attempts to make this more robust by allowing
dev_pm_opp_get_opp_table() to return an error pointer. Fixing all
the usages is trivial because the function is usually used indirectly
through another helper (e.g. dev_pm_opp_set_supported_hw() above).
These other helpers already return an error pointer.
The example above then works correctly because set_supported_hw() will
return -EPROBE_DEFER, and qcom-cpufreq-nvmem.c already propagates that
error. It should also be possible to remove the remaining usages of
"dev_pm_opp_of_find_icc_paths(dev, NULL)" from other drivers as well.
Note that this commit currently only handles -EPROBE_DEFER for the
clock/interconnects within _allocate_opp_table(). Other errors are just
ignored as before. Eventually those should be propagated as well.
Signed-off-by: Stephan Gerhold <stephan@gerhold.net>
Acked-by: Krzysztof Kozlowski <krzk@kernel.org>
Reviewed-by: Ulf Hansson <ulf.hansson@linaro.org>
[ Viresh: skip checking return value of dev_pm_opp_get_opp_table() for
EPROBE_DEFER in domain.c, fix NULL return value and reorder
code a bit in core.c, and update exynos-asv.c ]
Signed-off-by: Viresh Kumar <viresh.kumar@linaro.org>
2020-07-27 09:30:46 +00:00
|
|
|
WARN_ON(IS_ERR(genpd->opp_table));
|
2018-04-05 10:23:34 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
genpd->provider = &np->fwnode;
|
|
|
|
genpd->has_provider = true;
|
2016-09-12 11:01:10 +00:00
|
|
|
}
|
|
|
|
|
2017-03-29 16:34:50 +00:00
|
|
|
ret = genpd_add_provider(np, data->xlate, data);
|
2016-09-12 11:01:12 +00:00
|
|
|
if (ret < 0)
|
|
|
|
goto error;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
error:
|
|
|
|
while (i--) {
|
2018-04-05 10:23:34 +00:00
|
|
|
genpd = data->domains[i];
|
|
|
|
|
|
|
|
if (!genpd)
|
2016-09-15 12:05:23 +00:00
|
|
|
continue;
|
2018-04-05 10:23:34 +00:00
|
|
|
|
|
|
|
genpd->provider = NULL;
|
|
|
|
genpd->has_provider = false;
|
|
|
|
|
2023-08-25 11:26:32 +00:00
|
|
|
if (!genpd_is_opp_table_fw(genpd) && genpd->set_performance_state) {
|
2018-11-02 05:48:08 +00:00
|
|
|
dev_pm_opp_put_opp_table(genpd->opp_table);
|
2018-04-05 10:23:34 +00:00
|
|
|
dev_pm_opp_of_remove_table(&genpd->dev);
|
2018-11-02 05:48:08 +00:00
|
|
|
}
|
2016-09-12 11:01:12 +00:00
|
|
|
}
|
2016-09-12 11:01:10 +00:00
|
|
|
|
|
|
|
return ret;
|
2016-09-12 11:01:09 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(of_genpd_add_provider_onecell);
|
2014-09-19 18:27:36 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* of_genpd_del_provider() - Remove a previously registered PM domain provider
|
|
|
|
* @np: Device node pointer associated with the PM domain provider
|
|
|
|
*/
|
|
|
|
void of_genpd_del_provider(struct device_node *np)
|
|
|
|
{
|
2017-06-28 14:56:19 +00:00
|
|
|
struct of_genpd_provider *cp, *tmp;
|
2016-09-12 11:01:12 +00:00
|
|
|
struct generic_pm_domain *gpd;
|
2014-09-19 18:27:36 +00:00
|
|
|
|
2016-09-12 11:01:12 +00:00
|
|
|
mutex_lock(&gpd_list_lock);
|
2014-09-19 18:27:36 +00:00
|
|
|
mutex_lock(&of_genpd_mutex);
|
2017-06-28 14:56:19 +00:00
|
|
|
list_for_each_entry_safe(cp, tmp, &of_genpd_providers, link) {
|
2014-09-19 18:27:36 +00:00
|
|
|
if (cp->node == np) {
|
2016-09-12 11:01:12 +00:00
|
|
|
/*
|
|
|
|
* For each PM domain associated with the
|
|
|
|
* provider, set the 'has_provider' to false
|
|
|
|
* so that the PM domain can be safely removed.
|
|
|
|
*/
|
2018-04-05 10:23:34 +00:00
|
|
|
list_for_each_entry(gpd, &gpd_list, gpd_list_node) {
|
|
|
|
if (gpd->provider == &np->fwnode) {
|
2016-09-12 11:01:12 +00:00
|
|
|
gpd->has_provider = false;
|
|
|
|
|
2023-08-25 11:26:32 +00:00
|
|
|
if (genpd_is_opp_table_fw(gpd) || !gpd->set_performance_state)
|
2018-04-05 10:23:34 +00:00
|
|
|
continue;
|
|
|
|
|
2018-11-02 05:48:08 +00:00
|
|
|
dev_pm_opp_put_opp_table(gpd->opp_table);
|
2018-04-05 10:23:34 +00:00
|
|
|
dev_pm_opp_of_remove_table(&gpd->dev);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-02-05 22:26:43 +00:00
|
|
|
fwnode_dev_initialized(&cp->node->fwnode, false);
|
2014-09-19 18:27:36 +00:00
|
|
|
list_del(&cp->link);
|
|
|
|
of_node_put(cp->node);
|
|
|
|
kfree(cp);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
mutex_unlock(&of_genpd_mutex);
|
2016-09-12 11:01:12 +00:00
|
|
|
mutex_unlock(&gpd_list_lock);
|
2014-09-19 18:27:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(of_genpd_del_provider);
|
|
|
|
|
|
|
|
/**
|
2016-09-12 11:01:08 +00:00
|
|
|
* genpd_get_from_provider() - Look-up PM domain
|
2014-09-19 18:27:36 +00:00
|
|
|
* @genpdspec: OF phandle args to use for look-up
|
|
|
|
*
|
|
|
|
* Looks for a PM domain provider under the node specified by @genpdspec and if
|
|
|
|
* found, uses xlate function of the provider to map phandle args to a PM
|
|
|
|
* domain.
|
|
|
|
*
|
|
|
|
* Returns a valid pointer to struct generic_pm_domain on success or ERR_PTR()
|
|
|
|
* on failure.
|
|
|
|
*/
|
2016-09-12 11:01:08 +00:00
|
|
|
static struct generic_pm_domain *genpd_get_from_provider(
|
2024-02-08 20:28:22 +00:00
|
|
|
const struct of_phandle_args *genpdspec)
|
2014-09-19 18:27:36 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = ERR_PTR(-ENOENT);
|
|
|
|
struct of_genpd_provider *provider;
|
|
|
|
|
2016-03-04 10:55:15 +00:00
|
|
|
if (!genpdspec)
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
2014-09-19 18:27:36 +00:00
|
|
|
mutex_lock(&of_genpd_mutex);
|
|
|
|
|
|
|
|
/* Check if we have such a provider in our array */
|
|
|
|
list_for_each_entry(provider, &of_genpd_providers, link) {
|
|
|
|
if (provider->node == genpdspec->np)
|
|
|
|
genpd = provider->xlate(genpdspec, provider->data);
|
|
|
|
if (!IS_ERR(genpd))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
mutex_unlock(&of_genpd_mutex);
|
|
|
|
|
|
|
|
return genpd;
|
|
|
|
}
|
|
|
|
|
2016-09-12 11:01:05 +00:00
|
|
|
/**
|
|
|
|
* of_genpd_add_device() - Add a device to an I/O PM domain
|
|
|
|
* @genpdspec: OF phandle args to use for look-up PM domain
|
|
|
|
* @dev: Device to be added.
|
|
|
|
*
|
|
|
|
* Looks-up an I/O PM domain based upon phandle args provided and adds
|
|
|
|
* the device to the PM domain. Returns a negative error code on failure.
|
|
|
|
*/
|
2024-02-08 20:28:22 +00:00
|
|
|
int of_genpd_add_device(const struct of_phandle_args *genpdspec, struct device *dev)
|
2016-09-12 11:01:05 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd;
|
2016-09-12 11:01:11 +00:00
|
|
|
int ret;
|
|
|
|
|
2023-05-30 09:55:36 +00:00
|
|
|
if (!dev)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2016-09-12 11:01:11 +00:00
|
|
|
mutex_lock(&gpd_list_lock);
|
2016-09-12 11:01:05 +00:00
|
|
|
|
2016-09-12 11:01:08 +00:00
|
|
|
genpd = genpd_get_from_provider(genpdspec);
|
2016-09-12 11:01:11 +00:00
|
|
|
if (IS_ERR(genpd)) {
|
|
|
|
ret = PTR_ERR(genpd);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2019-04-25 09:04:13 +00:00
|
|
|
ret = genpd_add_device(genpd, dev, dev);
|
2016-09-12 11:01:05 +00:00
|
|
|
|
2016-09-12 11:01:11 +00:00
|
|
|
out:
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
|
|
|
return ret;
|
2016-09-12 11:01:05 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(of_genpd_add_device);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* of_genpd_add_subdomain - Add a subdomain to an I/O PM domain.
|
|
|
|
* @parent_spec: OF phandle args to use for parent PM domain look-up
|
|
|
|
* @subdomain_spec: OF phandle args to use for subdomain look-up
|
|
|
|
*
|
|
|
|
* Looks-up a parent PM domain and subdomain based upon phandle args
|
|
|
|
* provided and adds the subdomain to the parent PM domain. Returns a
|
|
|
|
* negative error code on failure.
|
|
|
|
*/
|
2024-02-08 20:28:22 +00:00
|
|
|
int of_genpd_add_subdomain(const struct of_phandle_args *parent_spec,
|
|
|
|
const struct of_phandle_args *subdomain_spec)
|
2016-09-12 11:01:05 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *parent, *subdomain;
|
2016-09-12 11:01:11 +00:00
|
|
|
int ret;
|
|
|
|
|
|
|
|
mutex_lock(&gpd_list_lock);
|
2016-09-12 11:01:05 +00:00
|
|
|
|
2016-09-12 11:01:08 +00:00
|
|
|
parent = genpd_get_from_provider(parent_spec);
|
2016-09-12 11:01:11 +00:00
|
|
|
if (IS_ERR(parent)) {
|
|
|
|
ret = PTR_ERR(parent);
|
|
|
|
goto out;
|
|
|
|
}
|
2016-09-12 11:01:05 +00:00
|
|
|
|
2016-09-12 11:01:08 +00:00
|
|
|
subdomain = genpd_get_from_provider(subdomain_spec);
|
2016-09-12 11:01:11 +00:00
|
|
|
if (IS_ERR(subdomain)) {
|
|
|
|
ret = PTR_ERR(subdomain);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = genpd_add_subdomain(parent, subdomain);
|
2016-09-12 11:01:05 +00:00
|
|
|
|
2016-09-12 11:01:11 +00:00
|
|
|
out:
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
2021-01-20 21:12:31 +00:00
|
|
|
return ret == -ENOENT ? -EPROBE_DEFER : ret;
|
2016-09-12 11:01:05 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(of_genpd_add_subdomain);
|
|
|
|
|
2019-12-30 12:59:30 +00:00
|
|
|
/**
|
|
|
|
* of_genpd_remove_subdomain - Remove a subdomain from an I/O PM domain.
|
|
|
|
* @parent_spec: OF phandle args to use for parent PM domain look-up
|
|
|
|
* @subdomain_spec: OF phandle args to use for subdomain look-up
|
|
|
|
*
|
|
|
|
* Looks-up a parent PM domain and subdomain based upon phandle args
|
|
|
|
* provided and removes the subdomain from the parent PM domain. Returns a
|
|
|
|
* negative error code on failure.
|
|
|
|
*/
|
2024-02-08 20:28:22 +00:00
|
|
|
int of_genpd_remove_subdomain(const struct of_phandle_args *parent_spec,
|
|
|
|
const struct of_phandle_args *subdomain_spec)
|
2019-12-30 12:59:30 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *parent, *subdomain;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
mutex_lock(&gpd_list_lock);
|
|
|
|
|
|
|
|
parent = genpd_get_from_provider(parent_spec);
|
|
|
|
if (IS_ERR(parent)) {
|
|
|
|
ret = PTR_ERR(parent);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
subdomain = genpd_get_from_provider(subdomain_spec);
|
|
|
|
if (IS_ERR(subdomain)) {
|
|
|
|
ret = PTR_ERR(subdomain);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = pm_genpd_remove_subdomain(parent, subdomain);
|
|
|
|
|
|
|
|
out:
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(of_genpd_remove_subdomain);
|
|
|
|
|
2016-09-12 11:01:14 +00:00
|
|
|
/**
|
|
|
|
* of_genpd_remove_last - Remove the last PM domain registered for a provider
|
2021-05-12 07:25:15 +00:00
|
|
|
* @np: Pointer to device node associated with provider
|
2016-09-12 11:01:14 +00:00
|
|
|
*
|
|
|
|
* Find the last PM domain that was added by a particular provider and
|
|
|
|
* remove this PM domain from the list of PM domains. The provider is
|
|
|
|
* identified by the 'provider' device structure that is passed. The PM
|
|
|
|
* domain will only be removed, if the provider associated with domain
|
|
|
|
* has been removed.
|
|
|
|
*
|
|
|
|
* Returns a valid pointer to struct generic_pm_domain on success or
|
|
|
|
* ERR_PTR() on failure.
|
|
|
|
*/
|
|
|
|
struct generic_pm_domain *of_genpd_remove_last(struct device_node *np)
|
|
|
|
{
|
2017-06-28 14:56:20 +00:00
|
|
|
struct generic_pm_domain *gpd, *tmp, *genpd = ERR_PTR(-ENOENT);
|
2016-09-12 11:01:14 +00:00
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (IS_ERR_OR_NULL(np))
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
|
|
|
mutex_lock(&gpd_list_lock);
|
2017-06-28 14:56:20 +00:00
|
|
|
list_for_each_entry_safe(gpd, tmp, &gpd_list, gpd_list_node) {
|
2016-09-12 11:01:14 +00:00
|
|
|
if (gpd->provider == &np->fwnode) {
|
|
|
|
ret = genpd_remove(gpd);
|
|
|
|
genpd = ret ? ERR_PTR(ret) : gpd;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
|
|
|
return genpd;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(of_genpd_remove_last);
|
|
|
|
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
static void genpd_release_dev(struct device *dev)
|
|
|
|
{
|
2019-04-18 10:27:56 +00:00
|
|
|
of_node_put(dev->of_node);
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
kfree(dev);
|
|
|
|
}
|
|
|
|
|
2023-12-19 15:35:09 +00:00
|
|
|
static const struct bus_type genpd_bus_type = {
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
.name = "genpd",
|
|
|
|
};
|
|
|
|
|
2014-09-19 18:27:36 +00:00
|
|
|
/**
|
|
|
|
* genpd_dev_pm_detach - Detach a device from its PM domain.
|
2015-08-27 09:17:00 +00:00
|
|
|
* @dev: Device to detach.
|
2014-09-19 18:27:36 +00:00
|
|
|
* @power_off: Currently not used
|
|
|
|
*
|
|
|
|
* Try to locate a corresponding generic PM domain, which the device was
|
|
|
|
* attached to previously. If such is found, the device is detached from it.
|
|
|
|
*/
|
|
|
|
static void genpd_dev_pm_detach(struct device *dev, bool power_off)
|
|
|
|
{
|
2015-03-20 17:20:33 +00:00
|
|
|
struct generic_pm_domain *pd;
|
2015-06-26 09:14:14 +00:00
|
|
|
unsigned int i;
|
2014-09-19 18:27:36 +00:00
|
|
|
int ret = 0;
|
|
|
|
|
2016-09-21 13:38:50 +00:00
|
|
|
pd = dev_to_genpd(dev);
|
|
|
|
if (IS_ERR(pd))
|
2014-09-19 18:27:36 +00:00
|
|
|
return;
|
|
|
|
|
|
|
|
dev_dbg(dev, "removing from PM domain %s\n", pd->name);
|
|
|
|
|
2021-08-12 11:27:21 +00:00
|
|
|
/* Drop the default performance state */
|
|
|
|
if (dev_gpd_data(dev)->default_pstate) {
|
|
|
|
dev_pm_genpd_set_performance_state(dev, 0);
|
|
|
|
dev_gpd_data(dev)->default_pstate = 0;
|
|
|
|
}
|
|
|
|
|
2015-06-26 09:14:14 +00:00
|
|
|
for (i = 1; i < GENPD_RETRY_MAX_MS; i <<= 1) {
|
2016-09-21 13:38:50 +00:00
|
|
|
ret = genpd_remove_device(pd, dev);
|
2014-09-19 18:27:36 +00:00
|
|
|
if (ret != -EAGAIN)
|
|
|
|
break;
|
2015-06-26 09:14:14 +00:00
|
|
|
|
|
|
|
mdelay(i);
|
2014-09-19 18:27:36 +00:00
|
|
|
cond_resched();
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ret < 0) {
|
|
|
|
dev_err(dev, "failed to remove from PM domain %s: %d",
|
|
|
|
pd->name, ret);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Check if PM domain can be powered off after removing this device. */
|
|
|
|
genpd_queue_power_off_work(pd);
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
|
|
|
|
/* Unregister the device if it was created by genpd. */
|
|
|
|
if (dev->bus == &genpd_bus_type)
|
|
|
|
device_unregister(dev);
|
2014-09-19 18:27:36 +00:00
|
|
|
}
|
|
|
|
|
2015-03-20 14:55:12 +00:00
|
|
|
static void genpd_dev_pm_sync(struct device *dev)
|
|
|
|
{
|
|
|
|
struct generic_pm_domain *pd;
|
|
|
|
|
|
|
|
pd = dev_to_genpd(dev);
|
|
|
|
if (IS_ERR(pd))
|
|
|
|
return;
|
|
|
|
|
|
|
|
genpd_queue_power_off_work(pd);
|
|
|
|
}
|
|
|
|
|
2019-04-25 09:04:10 +00:00
|
|
|
static int __genpd_dev_pm_attach(struct device *dev, struct device *base_dev,
|
|
|
|
unsigned int index, bool power_on)
|
2014-09-19 18:27:36 +00:00
|
|
|
{
|
|
|
|
struct of_phandle_args pd_args;
|
|
|
|
struct generic_pm_domain *pd;
|
2021-08-12 11:27:21 +00:00
|
|
|
int pstate;
|
2014-09-19 18:27:36 +00:00
|
|
|
int ret;
|
|
|
|
|
2019-04-18 10:27:56 +00:00
|
|
|
ret = of_parse_phandle_with_args(dev->of_node, "power-domains",
|
2018-05-31 10:59:57 +00:00
|
|
|
"#power-domain-cells", index, &pd_args);
|
2017-11-30 11:54:28 +00:00
|
|
|
if (ret < 0)
|
2018-05-31 10:59:56 +00:00
|
|
|
return ret;
|
2014-09-19 18:27:36 +00:00
|
|
|
|
2016-09-12 11:01:11 +00:00
|
|
|
mutex_lock(&gpd_list_lock);
|
2016-09-12 11:01:08 +00:00
|
|
|
pd = genpd_get_from_provider(&pd_args);
|
2015-12-01 17:39:31 +00:00
|
|
|
of_node_put(pd_args.np);
|
2014-09-19 18:27:36 +00:00
|
|
|
if (IS_ERR(pd)) {
|
2016-09-12 11:01:11 +00:00
|
|
|
mutex_unlock(&gpd_list_lock);
|
2014-09-19 18:27:36 +00:00
|
|
|
dev_dbg(dev, "%s() failed to find PM domain: %ld\n",
|
|
|
|
__func__, PTR_ERR(pd));
|
2022-08-19 22:16:13 +00:00
|
|
|
return driver_deferred_probe_check_state(base_dev);
|
2014-09-19 18:27:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
dev_dbg(dev, "adding to PM domain %s\n", pd->name);
|
|
|
|
|
2019-04-25 09:04:13 +00:00
|
|
|
ret = genpd_add_device(pd, dev, base_dev);
|
2016-09-12 11:01:11 +00:00
|
|
|
mutex_unlock(&gpd_list_lock);
|
2014-09-19 18:27:36 +00:00
|
|
|
|
2022-02-23 08:03:23 +00:00
|
|
|
if (ret < 0)
|
|
|
|
return dev_err_probe(dev, ret, "failed to add to PM domain %s\n", pd->name);
|
2014-09-19 18:27:36 +00:00
|
|
|
|
|
|
|
dev->pm_domain->detach = genpd_dev_pm_detach;
|
2015-03-20 14:55:12 +00:00
|
|
|
dev->pm_domain->sync = genpd_dev_pm_sync;
|
2014-09-19 18:27:36 +00:00
|
|
|
|
2021-08-12 11:27:21 +00:00
|
|
|
/* Set the default performance state */
|
|
|
|
pstate = of_get_required_opp_performance_state(dev->of_node, index);
|
2021-08-24 15:23:38 +00:00
|
|
|
if (pstate < 0 && pstate != -ENODEV && pstate != -EOPNOTSUPP) {
|
2021-08-12 11:27:21 +00:00
|
|
|
ret = pstate;
|
|
|
|
goto err;
|
|
|
|
} else if (pstate > 0) {
|
|
|
|
ret = dev_pm_genpd_set_performance_state(dev, pstate);
|
|
|
|
if (ret)
|
|
|
|
goto err;
|
|
|
|
dev_gpd_data(dev)->default_pstate = pstate;
|
|
|
|
}
|
PM: domains: Reverse the order of performance and enabling ops
The ->set_performance_state() needs to be called before ->power_on()
when a genpd is powered on, and after ->power_off() when a genpd is
powered off. Do this in order to let the provider know to which
performance state to power on the genpd, on the power on sequence, and
also to maintain the performance for that genpd until after powering off,
on power off sequence.
There is no scenario where a consumer would need its genpd enabled and
then its performance state increased. Instead, in every scenario, the
consumer needs the genpd to be enabled from the start at a specific
performance state.
And same logic applies to the powering down. No consumer would need its
genpd performance state dropped right before powering down.
Now, there are currently two vendors which use ->set_performance_state()
in their genpd providers. One of them is Tegra, but the only genpd provider
(PMC) that makes use of ->set_performance_state() doesn't implement the
->power_on() or ->power_off(), and so it will not be affected by the ops
reversal.
The other vendor that uses it is Qualcomm, in multiple genpd providers
actually (RPM, RPMh and CPR). But all Qualcomm genpd providers that make
use of ->set_performance_state() need the order between enabling ops and
the performance setting op to be reversed. And the reason for that is that
it currently translates into two different voltages in order to power on
a genpd to a specific performance state. Basically, ->power_on() switches
to the minimum (enabling) voltage for that genpd, and then
->set_performance_state() sets it to the voltage level required by the
consumer.
By reversing the call order, we rely on the provider to know what to do
on each call, but most popular usecase is to cache the performance state
and postpone the voltage setting until the ->power_on() gets called.
As for the reason of still needing the ->power_on() and ->power_off() for a
provider which could get away with just having ->set_performance_state()
implemented, there are consumers that do not (nor should) provide an
opp-table. For those consumers, ->set_performance_state() will not be
called, and so they will enable the genpd to its minimum performance state
by a ->power_on() call. Same logic goes for the disabling.
Signed-off-by: Abel Vesa <abel.vesa@linaro.org>
Reviewed-by: Ulf Hansson <ulf.hansson@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2022-11-15 21:25:43 +00:00
|
|
|
|
|
|
|
if (power_on) {
|
|
|
|
genpd_lock(pd);
|
|
|
|
ret = genpd_power_on(pd, 0);
|
|
|
|
genpd_unlock(pd);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ret) {
|
|
|
|
/* Drop the default performance state */
|
|
|
|
if (dev_gpd_data(dev)->default_pstate) {
|
|
|
|
dev_pm_genpd_set_performance_state(dev, 0);
|
|
|
|
dev_gpd_data(dev)->default_pstate = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
genpd_remove_device(pd, dev);
|
|
|
|
return -EPROBE_DEFER;
|
|
|
|
}
|
|
|
|
|
2021-08-12 11:27:21 +00:00
|
|
|
return 1;
|
|
|
|
|
|
|
|
err:
|
|
|
|
dev_err(dev, "failed to set required performance state for power-domain %s: %d\n",
|
|
|
|
pd->name, ret);
|
|
|
|
genpd_remove_device(pd, dev);
|
|
|
|
return ret;
|
2014-09-19 18:27:36 +00:00
|
|
|
}
|
2018-05-31 10:59:57 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* genpd_dev_pm_attach - Attach a device to its PM domain using DT.
|
|
|
|
* @dev: Device to attach.
|
|
|
|
*
|
|
|
|
* Parse device's OF node to find a PM domain specifier. If such is found,
|
|
|
|
* attaches the device to retrieved pm_domain ops.
|
|
|
|
*
|
|
|
|
* Returns 1 on successfully attached PM domain, 0 when the device don't need a
|
|
|
|
* PM domain or when multiple power-domains exists for it, else a negative error
|
|
|
|
* code. Note that if a power-domain exists for the device, but it cannot be
|
|
|
|
* found or turned on, then return -EPROBE_DEFER to ensure that the device is
|
|
|
|
* not probed and to re-try again later.
|
|
|
|
*/
|
|
|
|
int genpd_dev_pm_attach(struct device *dev)
|
|
|
|
{
|
|
|
|
if (!dev->of_node)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Devices with multiple PM domains must be attached separately, as we
|
|
|
|
* can only attach one PM domain per device.
|
|
|
|
*/
|
|
|
|
if (of_count_phandle_with_args(dev->of_node, "power-domains",
|
|
|
|
"#power-domain-cells") != 1)
|
|
|
|
return 0;
|
|
|
|
|
2019-04-25 09:04:10 +00:00
|
|
|
return __genpd_dev_pm_attach(dev, dev, 0, true);
|
2018-05-31 10:59:57 +00:00
|
|
|
}
|
2014-09-19 18:27:36 +00:00
|
|
|
EXPORT_SYMBOL_GPL(genpd_dev_pm_attach);
|
2016-10-14 17:47:51 +00:00
|
|
|
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
/**
|
|
|
|
* genpd_dev_pm_attach_by_id - Associate a device with one of its PM domains.
|
|
|
|
* @dev: The device used to lookup the PM domain.
|
|
|
|
* @index: The index of the PM domain.
|
|
|
|
*
|
|
|
|
* Parse device's OF node to find a PM domain specifier at the provided @index.
|
|
|
|
* If such is found, creates a virtual device and attaches it to the retrieved
|
|
|
|
* pm_domain ops. To deal with detaching of the virtual device, the ->detach()
|
|
|
|
* callback in the struct dev_pm_domain are assigned to genpd_dev_pm_detach().
|
|
|
|
*
|
|
|
|
* Returns the created virtual device if successfully attached PM domain, NULL
|
|
|
|
* when the device don't need a PM domain, else an ERR_PTR() in case of
|
|
|
|
* failures. If a power-domain exists for the device, but cannot be found or
|
|
|
|
* turned on, then ERR_PTR(-EPROBE_DEFER) is returned to ensure that the device
|
|
|
|
* is not probed and to re-try again later.
|
|
|
|
*/
|
|
|
|
struct device *genpd_dev_pm_attach_by_id(struct device *dev,
|
|
|
|
unsigned int index)
|
|
|
|
{
|
2018-10-25 03:37:38 +00:00
|
|
|
struct device *virt_dev;
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
int num_domains;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (!dev->of_node)
|
|
|
|
return NULL;
|
|
|
|
|
2019-04-18 10:27:57 +00:00
|
|
|
/* Verify that the index is within a valid range. */
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
num_domains = of_count_phandle_with_args(dev->of_node, "power-domains",
|
|
|
|
"#power-domain-cells");
|
2019-04-18 10:27:57 +00:00
|
|
|
if (index >= num_domains)
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
return NULL;
|
|
|
|
|
|
|
|
/* Allocate and register device on the genpd bus. */
|
2018-10-25 03:37:38 +00:00
|
|
|
virt_dev = kzalloc(sizeof(*virt_dev), GFP_KERNEL);
|
|
|
|
if (!virt_dev)
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
2018-10-25 03:37:38 +00:00
|
|
|
dev_set_name(virt_dev, "genpd:%u:%s", index, dev_name(dev));
|
|
|
|
virt_dev->bus = &genpd_bus_type;
|
|
|
|
virt_dev->release = genpd_release_dev;
|
2019-04-18 10:27:56 +00:00
|
|
|
virt_dev->of_node = of_node_get(dev->of_node);
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
|
2018-10-25 03:37:38 +00:00
|
|
|
ret = device_register(virt_dev);
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
if (ret) {
|
2019-04-18 10:27:55 +00:00
|
|
|
put_device(virt_dev);
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
return ERR_PTR(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Try to attach the device to the PM domain at the specified index. */
|
2019-04-25 09:04:10 +00:00
|
|
|
ret = __genpd_dev_pm_attach(virt_dev, dev, index, false);
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
if (ret < 1) {
|
2018-10-25 03:37:38 +00:00
|
|
|
device_unregister(virt_dev);
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
return ret ? ERR_PTR(ret) : NULL;
|
|
|
|
}
|
|
|
|
|
2018-10-25 03:37:38 +00:00
|
|
|
pm_runtime_enable(virt_dev);
|
|
|
|
genpd_queue_power_off_work(dev_to_genpd(virt_dev));
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
|
2018-10-25 03:37:38 +00:00
|
|
|
return virt_dev;
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(genpd_dev_pm_attach_by_id);
|
|
|
|
|
2018-06-29 11:04:31 +00:00
|
|
|
/**
|
|
|
|
* genpd_dev_pm_attach_by_name - Associate a device with one of its PM domains.
|
|
|
|
* @dev: The device used to lookup the PM domain.
|
|
|
|
* @name: The name of the PM domain.
|
|
|
|
*
|
|
|
|
* Parse device's OF node to find a PM domain specifier using the
|
|
|
|
* power-domain-names DT property. For further description see
|
|
|
|
* genpd_dev_pm_attach_by_id().
|
|
|
|
*/
|
2019-02-14 18:12:48 +00:00
|
|
|
struct device *genpd_dev_pm_attach_by_name(struct device *dev, const char *name)
|
2018-06-29 11:04:31 +00:00
|
|
|
{
|
|
|
|
int index;
|
|
|
|
|
|
|
|
if (!dev->of_node)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
index = of_property_match_string(dev->of_node, "power-domain-names",
|
|
|
|
name);
|
|
|
|
if (index < 0)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
return genpd_dev_pm_attach_by_id(dev, index);
|
|
|
|
}
|
|
|
|
|
2016-10-14 17:47:51 +00:00
|
|
|
static const struct of_device_id idle_state_match[] = {
|
2016-11-03 21:54:35 +00:00
|
|
|
{ .compatible = "domain-idle-state", },
|
2016-10-14 17:47:51 +00:00
|
|
|
{ }
|
|
|
|
};
|
|
|
|
|
|
|
|
static int genpd_parse_state(struct genpd_power_state *genpd_state,
|
|
|
|
struct device_node *state_node)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
u32 residency;
|
|
|
|
u32 entry_latency, exit_latency;
|
|
|
|
|
|
|
|
err = of_property_read_u32(state_node, "entry-latency-us",
|
|
|
|
&entry_latency);
|
|
|
|
if (err) {
|
2017-07-18 21:42:50 +00:00
|
|
|
pr_debug(" * %pOF missing entry-latency-us property\n",
|
2019-03-04 17:14:38 +00:00
|
|
|
state_node);
|
2016-10-14 17:47:51 +00:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = of_property_read_u32(state_node, "exit-latency-us",
|
|
|
|
&exit_latency);
|
|
|
|
if (err) {
|
2017-07-18 21:42:50 +00:00
|
|
|
pr_debug(" * %pOF missing exit-latency-us property\n",
|
2019-03-04 17:14:38 +00:00
|
|
|
state_node);
|
2016-10-14 17:47:51 +00:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = of_property_read_u32(state_node, "min-residency-us", &residency);
|
|
|
|
if (!err)
|
2023-04-18 13:07:43 +00:00
|
|
|
genpd_state->residency_ns = 1000LL * residency;
|
2016-10-14 17:47:51 +00:00
|
|
|
|
2023-04-18 13:07:43 +00:00
|
|
|
genpd_state->power_on_latency_ns = 1000LL * exit_latency;
|
|
|
|
genpd_state->power_off_latency_ns = 1000LL * entry_latency;
|
2016-10-14 17:47:52 +00:00
|
|
|
genpd_state->fwnode = &state_node->fwnode;
|
2016-10-14 17:47:51 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-01-23 20:43:08 +00:00
|
|
|
static int genpd_iterate_idle_states(struct device_node *dn,
|
|
|
|
struct genpd_power_state *states)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
struct of_phandle_iterator it;
|
|
|
|
struct device_node *np;
|
|
|
|
int i = 0;
|
|
|
|
|
|
|
|
ret = of_count_phandle_with_args(dn, "domain-idle-states", NULL);
|
|
|
|
if (ret <= 0)
|
2020-03-10 10:40:23 +00:00
|
|
|
return ret == -ENOENT ? 0 : ret;
|
2018-01-23 20:43:08 +00:00
|
|
|
|
|
|
|
/* Loop over the phandles until all the requested entry is found */
|
|
|
|
of_for_each_phandle(&it, ret, dn, "domain-idle-states", NULL, 0) {
|
|
|
|
np = it.node;
|
|
|
|
if (!of_match_node(idle_state_match, np))
|
|
|
|
continue;
|
2022-10-25 12:34:32 +00:00
|
|
|
|
|
|
|
if (!of_device_is_available(np))
|
|
|
|
continue;
|
|
|
|
|
2018-01-23 20:43:08 +00:00
|
|
|
if (states) {
|
|
|
|
ret = genpd_parse_state(&states[i], np);
|
|
|
|
if (ret) {
|
|
|
|
pr_err("Parsing idle state node %pOF failed with err %d\n",
|
|
|
|
np, ret);
|
|
|
|
of_node_put(np);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
i++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return i;
|
|
|
|
}
|
|
|
|
|
2016-10-14 17:47:51 +00:00
|
|
|
/**
|
|
|
|
* of_genpd_parse_idle_states: Return array of idle states for the genpd.
|
|
|
|
*
|
|
|
|
* @dn: The genpd device node
|
|
|
|
* @states: The pointer to which the state array will be saved.
|
|
|
|
* @n: The count of elements in the array returned from this function.
|
|
|
|
*
|
|
|
|
* Returns the device states parsed from the OF node. The memory for the states
|
|
|
|
* is allocated by this function and is the responsibility of the caller to
|
2018-10-03 14:38:14 +00:00
|
|
|
* free the memory after use. If any or zero compatible domain idle states is
|
|
|
|
* found it returns 0 and in case of errors, a negative error code is returned.
|
2016-10-14 17:47:51 +00:00
|
|
|
*/
|
|
|
|
int of_genpd_parse_idle_states(struct device_node *dn,
|
|
|
|
struct genpd_power_state **states, int *n)
|
|
|
|
{
|
|
|
|
struct genpd_power_state *st;
|
2018-01-23 20:43:08 +00:00
|
|
|
int ret;
|
2016-10-14 17:47:51 +00:00
|
|
|
|
2018-01-23 20:43:08 +00:00
|
|
|
ret = genpd_iterate_idle_states(dn, NULL);
|
2018-10-03 14:38:14 +00:00
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
if (!ret) {
|
|
|
|
*states = NULL;
|
|
|
|
*n = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
2016-10-14 17:47:51 +00:00
|
|
|
|
2018-01-23 20:43:08 +00:00
|
|
|
st = kcalloc(ret, sizeof(*st), GFP_KERNEL);
|
2016-10-14 17:47:51 +00:00
|
|
|
if (!st)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2018-01-23 20:43:08 +00:00
|
|
|
ret = genpd_iterate_idle_states(dn, st);
|
|
|
|
if (ret <= 0) {
|
|
|
|
kfree(st);
|
|
|
|
return ret < 0 ? ret : -EINVAL;
|
2016-10-14 17:47:51 +00:00
|
|
|
}
|
|
|
|
|
2018-01-23 20:43:08 +00:00
|
|
|
*states = st;
|
|
|
|
*n = ret;
|
2016-10-14 17:47:51 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(of_genpd_parse_idle_states);
|
|
|
|
|
PM / Domains: Add support for multi PM domains per device to genpd
To support devices being partitioned across multiple PM domains, let's
begin with extending genpd to cope with these kind of configurations.
Therefore, add a new exported function genpd_dev_pm_attach_by_id(), which
is similar to the existing genpd_dev_pm_attach(), but with the difference
that it allows its callers to provide an index to the PM domain that it
wants to attach.
Note that, genpd_dev_pm_attach_by_id() shall only be called by the driver
core / PM core, similar to how the existing dev_pm_domain_attach() makes
use of genpd_dev_pm_attach(). However, this is implemented by following
changes on top.
Because, only one PM domain can be attached per device, genpd needs to
create a virtual device that it can attach/detach instead. More precisely,
let the new function genpd_dev_pm_attach_by_id() register a virtual struct
device via calling device_register(). Then let it attach this device to the
corresponding PM domain, rather than the one that is provided by the
caller. The actual attaching is done via re-using the existing genpd OF
functions.
At successful attachment, genpd_dev_pm_attach_by_id() returns the created
virtual device, which allows the caller to operate on it to deal with power
management. Following changes on top, provides more details in this
regards.
To deal with detaching of a PM domain for the multiple PM domains case,
let's also extend the existing genpd_dev_pm_detach() function, to cover the
cleanup of the created virtual device, via make it call device_unregister()
on it. In this way, there is no need to introduce a new function to deal
with detach for the multiple PM domain case, but instead the existing one
is re-used.
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Jon Hunter <jonathanh@nvidia.com>
Tested-by: Jon Hunter <jonathanh@nvidia.com>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2018-05-31 10:59:58 +00:00
|
|
|
static int __init genpd_bus_init(void)
|
|
|
|
{
|
|
|
|
return bus_register(&genpd_bus_type);
|
|
|
|
}
|
|
|
|
core_initcall(genpd_bus_init);
|
|
|
|
|
2014-11-27 21:38:05 +00:00
|
|
|
#endif /* CONFIG_PM_GENERIC_DOMAINS_OF */
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
|
|
|
|
|
|
|
|
/*** debugfs support ***/
|
|
|
|
|
2016-08-11 10:40:05 +00:00
|
|
|
#ifdef CONFIG_DEBUG_FS
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
/*
|
|
|
|
* TODO: This function is a slightly modified version of rtpm_status_show
|
2014-11-27 21:38:05 +00:00
|
|
|
* from sysfs.c, so generalize it.
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
*/
|
|
|
|
static void rtpm_status_str(struct seq_file *s, struct device *dev)
|
|
|
|
{
|
|
|
|
static const char * const status_lookup[] = {
|
|
|
|
[RPM_ACTIVE] = "active",
|
|
|
|
[RPM_RESUMING] = "resuming",
|
|
|
|
[RPM_SUSPENDED] = "suspended",
|
|
|
|
[RPM_SUSPENDING] = "suspending"
|
|
|
|
};
|
|
|
|
const char *p = "";
|
|
|
|
|
|
|
|
if (dev->power.runtime_error)
|
|
|
|
p = "error";
|
|
|
|
else if (dev->power.disable_depth)
|
|
|
|
p = "unsupported";
|
|
|
|
else if (dev->power.runtime_status < ARRAY_SIZE(status_lookup))
|
|
|
|
p = status_lookup[dev->power.runtime_status];
|
|
|
|
else
|
|
|
|
WARN_ON(1);
|
|
|
|
|
pmdomain: core: Reduce debug summary table width
Commit 9094e53ff5c86ebe ("pmdomain: core: Use dev_name() instead of
kobject_get_path() in debugfs") severely shortened the names of devices
in a PM Domain. Now the most common format[1] consists of a 32-bit
unit-address (8 characters), followed by a dot and a node name (20
characters for "air-pollution-sensor" and "interrupt-controller", which
are the longest generic node names documented in the Devicetree
Specification), for a typical maximum of 29 characters.
This offers a good opportunity to reduce the table width of the debug
summary:
- Reduce the device name field width from 50 to 30 characters, which
matches the PM Domain name width,
- Reduce the large inter-column space between the "performance" and
"managed by" columns.
Visual impact:
- The "performance" column now starts at a position that is a
multiple of 16, just like the "status" and "children" columns,
- All of the "/device", "runtime status", and "managed by" columns are
now indented 4 characters more than the columns right above them,
- Everything fits in (one less than) 80 characters again ;-)
[1] Note that some device names (e.g. TI AM335x interconnect target
modules) do not follow this convention, and may be much longer, but
these didn't fit in the old 50-character column width either.
Signed-off-by: Geert Uytterhoeven <geert+renesas@glider.be>
Link: https://lore.kernel.org/r/f8e1821364b6d5d11350447c128f6d2b470f33fe.1725459707.git.geert+renesas@glider.be
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-09-04 14:30:48 +00:00
|
|
|
seq_printf(s, "%-26s ", p);
|
2021-01-20 21:12:32 +00:00
|
|
|
}
|
|
|
|
|
2024-09-04 14:30:47 +00:00
|
|
|
static void perf_status_str(struct seq_file *s, struct device *dev)
|
2024-06-24 04:48:06 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain_data *gpd_data;
|
|
|
|
|
|
|
|
gpd_data = to_gpd_data(dev->power.subsys_data->domain_data);
|
|
|
|
|
2024-09-04 14:30:47 +00:00
|
|
|
seq_printf(s, "%-10u ", gpd_data->performance_state);
|
2024-06-24 04:48:06 +00:00
|
|
|
}
|
|
|
|
|
2024-09-04 14:30:47 +00:00
|
|
|
static void mode_status_str(struct seq_file *s, struct device *dev)
|
2021-01-20 21:12:32 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain_data *gpd_data;
|
|
|
|
|
|
|
|
gpd_data = to_gpd_data(dev->power.subsys_data->domain_data);
|
2024-09-04 14:30:47 +00:00
|
|
|
|
pmdomain: core: Reduce debug summary table width
Commit 9094e53ff5c86ebe ("pmdomain: core: Use dev_name() instead of
kobject_get_path() in debugfs") severely shortened the names of devices
in a PM Domain. Now the most common format[1] consists of a 32-bit
unit-address (8 characters), followed by a dot and a node name (20
characters for "air-pollution-sensor" and "interrupt-controller", which
are the longest generic node names documented in the Devicetree
Specification), for a typical maximum of 29 characters.
This offers a good opportunity to reduce the table width of the debug
summary:
- Reduce the device name field width from 50 to 30 characters, which
matches the PM Domain name width,
- Reduce the large inter-column space between the "performance" and
"managed by" columns.
Visual impact:
- The "performance" column now starts at a position that is a
multiple of 16, just like the "status" and "children" columns,
- All of the "/device", "runtime status", and "managed by" columns are
now indented 4 characters more than the columns right above them,
- Everything fits in (one less than) 80 characters again ;-)
[1] Note that some device names (e.g. TI AM335x interconnect target
modules) do not follow this convention, and may be much longer, but
these didn't fit in the old 50-character column width either.
Signed-off-by: Geert Uytterhoeven <geert+renesas@glider.be>
Link: https://lore.kernel.org/r/f8e1821364b6d5d11350447c128f6d2b470f33fe.1725459707.git.geert+renesas@glider.be
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-09-04 14:30:48 +00:00
|
|
|
seq_printf(s, "%2s", gpd_data->hw_mode ? "HW" : "SW");
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
}
|
|
|
|
|
2017-10-06 07:02:06 +00:00
|
|
|
static int genpd_summary_one(struct seq_file *s,
|
|
|
|
struct generic_pm_domain *genpd)
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
{
|
|
|
|
static const char * const status_lookup[] = {
|
2020-09-24 11:04:47 +00:00
|
|
|
[GENPD_STATE_ON] = "on",
|
|
|
|
[GENPD_STATE_OFF] = "off"
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
};
|
|
|
|
struct pm_domain_data *pm_data;
|
|
|
|
struct gpd_link *link;
|
2016-02-23 16:49:17 +00:00
|
|
|
char state[16];
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
int ret;
|
|
|
|
|
2016-10-14 17:47:54 +00:00
|
|
|
ret = genpd_lock_interruptible(genpd);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
if (ret)
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
|
2015-03-02 19:24:28 +00:00
|
|
|
if (WARN_ON(genpd->status >= ARRAY_SIZE(status_lookup)))
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
goto exit;
|
2017-03-20 10:19:20 +00:00
|
|
|
if (!genpd_status_on(genpd))
|
2016-02-23 16:49:18 +00:00
|
|
|
snprintf(state, sizeof(state), "%s-%u",
|
2016-02-23 16:49:17 +00:00
|
|
|
status_lookup[genpd->status], genpd->state_idx);
|
2016-02-15 10:10:51 +00:00
|
|
|
else
|
2016-02-23 16:49:17 +00:00
|
|
|
snprintf(state, sizeof(state), "%s",
|
|
|
|
status_lookup[genpd->status]);
|
2024-10-30 12:55:10 +00:00
|
|
|
seq_printf(s, "%-30s %-30s %u", dev_name(&genpd->dev), state, genpd->performance_state);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Modifications on the list require holding locks on both
|
2020-07-08 23:32:13 +00:00
|
|
|
* parent and child, so we are safe.
|
2024-10-30 12:55:10 +00:00
|
|
|
* Also the device name is immutable.
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
*/
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry(link, &genpd->parent_links, parent_node) {
|
2021-01-20 21:12:32 +00:00
|
|
|
if (list_is_first(&link->parent_node, &genpd->parent_links))
|
|
|
|
seq_printf(s, "\n%48s", " ");
|
2020-07-08 23:32:13 +00:00
|
|
|
seq_printf(s, "%s", link->child->name);
|
|
|
|
if (!list_is_last(&link->parent_node, &genpd->parent_links))
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
seq_puts(s, ", ");
|
|
|
|
}
|
|
|
|
|
2015-03-02 19:24:28 +00:00
|
|
|
list_for_each_entry(pm_data, &genpd->dev_list, list_node) {
|
pmdomain: core: Reduce debug summary table width
Commit 9094e53ff5c86ebe ("pmdomain: core: Use dev_name() instead of
kobject_get_path() in debugfs") severely shortened the names of devices
in a PM Domain. Now the most common format[1] consists of a 32-bit
unit-address (8 characters), followed by a dot and a node name (20
characters for "air-pollution-sensor" and "interrupt-controller", which
are the longest generic node names documented in the Devicetree
Specification), for a typical maximum of 29 characters.
This offers a good opportunity to reduce the table width of the debug
summary:
- Reduce the device name field width from 50 to 30 characters, which
matches the PM Domain name width,
- Reduce the large inter-column space between the "performance" and
"managed by" columns.
Visual impact:
- The "performance" column now starts at a position that is a
multiple of 16, just like the "status" and "children" columns,
- All of the "/device", "runtime status", and "managed by" columns are
now indented 4 characters more than the columns right above them,
- Everything fits in (one less than) 80 characters again ;-)
[1] Note that some device names (e.g. TI AM335x interconnect target
modules) do not follow this convention, and may be much longer, but
these didn't fit in the old 50-character column width either.
Signed-off-by: Geert Uytterhoeven <geert+renesas@glider.be>
Link: https://lore.kernel.org/r/f8e1821364b6d5d11350447c128f6d2b470f33fe.1725459707.git.geert+renesas@glider.be
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-09-04 14:30:48 +00:00
|
|
|
seq_printf(s, "\n %-30s ", dev_name(pm_data->dev));
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
rtpm_status_str(s, pm_data->dev);
|
2021-01-20 21:12:32 +00:00
|
|
|
perf_status_str(s, pm_data->dev);
|
2024-06-24 04:48:06 +00:00
|
|
|
mode_status_str(s, pm_data->dev);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
seq_puts(s, "\n");
|
|
|
|
exit:
|
2016-10-14 17:47:54 +00:00
|
|
|
genpd_unlock(genpd);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-12-15 08:45:26 +00:00
|
|
|
static int summary_show(struct seq_file *s, void *data)
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
{
|
2015-03-02 19:24:28 +00:00
|
|
|
struct generic_pm_domain *genpd;
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
int ret = 0;
|
|
|
|
|
pmdomain: core: Reduce debug summary table width
Commit 9094e53ff5c86ebe ("pmdomain: core: Use dev_name() instead of
kobject_get_path() in debugfs") severely shortened the names of devices
in a PM Domain. Now the most common format[1] consists of a 32-bit
unit-address (8 characters), followed by a dot and a node name (20
characters for "air-pollution-sensor" and "interrupt-controller", which
are the longest generic node names documented in the Devicetree
Specification), for a typical maximum of 29 characters.
This offers a good opportunity to reduce the table width of the debug
summary:
- Reduce the device name field width from 50 to 30 characters, which
matches the PM Domain name width,
- Reduce the large inter-column space between the "performance" and
"managed by" columns.
Visual impact:
- The "performance" column now starts at a position that is a
multiple of 16, just like the "status" and "children" columns,
- All of the "/device", "runtime status", and "managed by" columns are
now indented 4 characters more than the columns right above them,
- Everything fits in (one less than) 80 characters again ;-)
[1] Note that some device names (e.g. TI AM335x interconnect target
modules) do not follow this convention, and may be much longer, but
these didn't fit in the old 50-character column width either.
Signed-off-by: Geert Uytterhoeven <geert+renesas@glider.be>
Link: https://lore.kernel.org/r/f8e1821364b6d5d11350447c128f6d2b470f33fe.1725459707.git.geert+renesas@glider.be
Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
2024-09-04 14:30:48 +00:00
|
|
|
seq_puts(s, "domain status children performance\n");
|
|
|
|
seq_puts(s, " /device runtime status managed by\n");
|
|
|
|
seq_puts(s, "------------------------------------------------------------------------------\n");
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
|
|
|
|
ret = mutex_lock_interruptible(&gpd_list_lock);
|
|
|
|
if (ret)
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
|
2015-03-02 19:24:28 +00:00
|
|
|
list_for_each_entry(genpd, &gpd_list, gpd_list_node) {
|
2017-10-06 07:02:06 +00:00
|
|
|
ret = genpd_summary_one(s, genpd);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
if (ret)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
mutex_unlock(&gpd_list_lock);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2018-12-15 08:45:26 +00:00
|
|
|
static int status_show(struct seq_file *s, void *data)
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
{
|
2017-07-14 17:10:16 +00:00
|
|
|
static const char * const status_lookup[] = {
|
2020-09-24 11:04:47 +00:00
|
|
|
[GENPD_STATE_ON] = "on",
|
|
|
|
[GENPD_STATE_OFF] = "off"
|
2017-07-14 17:10:16 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
struct generic_pm_domain *genpd = s->private;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
ret = genpd_lock_interruptible(genpd);
|
|
|
|
if (ret)
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
|
|
|
|
if (WARN_ON_ONCE(genpd->status >= ARRAY_SIZE(status_lookup)))
|
|
|
|
goto exit;
|
|
|
|
|
2020-09-24 11:04:47 +00:00
|
|
|
if (genpd->status == GENPD_STATE_OFF)
|
2017-07-14 17:10:16 +00:00
|
|
|
seq_printf(s, "%s-%u\n", status_lookup[genpd->status],
|
|
|
|
genpd->state_idx);
|
|
|
|
else
|
|
|
|
seq_printf(s, "%s\n", status_lookup[genpd->status]);
|
|
|
|
exit:
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
return ret;
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
}
|
|
|
|
|
2018-12-15 08:45:26 +00:00
|
|
|
static int sub_domains_show(struct seq_file *s, void *data)
|
2017-07-14 17:10:16 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = s->private;
|
|
|
|
struct gpd_link *link;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
ret = genpd_lock_interruptible(genpd);
|
|
|
|
if (ret)
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
|
2020-07-08 23:32:13 +00:00
|
|
|
list_for_each_entry(link, &genpd->parent_links, parent_node)
|
|
|
|
seq_printf(s, "%s\n", link->child->name);
|
2017-07-14 17:10:16 +00:00
|
|
|
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2018-12-15 08:45:26 +00:00
|
|
|
static int idle_states_show(struct seq_file *s, void *data)
|
2017-07-14 17:10:16 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = s->private;
|
2022-04-19 17:29:16 +00:00
|
|
|
u64 now, delta, idle_time = 0;
|
2017-07-14 17:10:16 +00:00
|
|
|
unsigned int i;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
ret = genpd_lock_interruptible(genpd);
|
|
|
|
if (ret)
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
|
2020-10-15 20:47:22 +00:00
|
|
|
seq_puts(s, "State Time Spent(ms) Usage Rejected\n");
|
2017-07-14 17:10:16 +00:00
|
|
|
|
|
|
|
for (i = 0; i < genpd->state_count; i++) {
|
2022-04-19 17:29:16 +00:00
|
|
|
idle_time += genpd->states[i].idle_time;
|
2017-07-14 17:10:16 +00:00
|
|
|
|
2022-04-19 17:29:16 +00:00
|
|
|
if (genpd->status == GENPD_STATE_OFF && genpd->state_idx == i) {
|
|
|
|
now = ktime_get_mono_fast_ns();
|
|
|
|
if (now > genpd->accounting_time) {
|
|
|
|
delta = now - genpd->accounting_time;
|
|
|
|
idle_time += delta;
|
|
|
|
}
|
|
|
|
}
|
2017-07-14 17:10:16 +00:00
|
|
|
|
2022-04-19 17:29:16 +00:00
|
|
|
do_div(idle_time, NSEC_PER_MSEC);
|
|
|
|
seq_printf(s, "S%-13i %-14llu %-14llu %llu\n", i, idle_time,
|
|
|
|
genpd->states[i].usage, genpd->states[i].rejected);
|
2017-07-14 17:10:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2018-12-15 08:45:26 +00:00
|
|
|
static int active_time_show(struct seq_file *s, void *data)
|
2017-07-14 17:10:16 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = s->private;
|
2022-04-19 17:29:16 +00:00
|
|
|
u64 now, on_time, delta = 0;
|
2017-07-14 17:10:16 +00:00
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
ret = genpd_lock_interruptible(genpd);
|
|
|
|
if (ret)
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
|
2022-04-19 17:29:16 +00:00
|
|
|
if (genpd->status == GENPD_STATE_ON) {
|
|
|
|
now = ktime_get_mono_fast_ns();
|
|
|
|
if (now > genpd->accounting_time)
|
|
|
|
delta = now - genpd->accounting_time;
|
|
|
|
}
|
2017-07-14 17:10:16 +00:00
|
|
|
|
2022-04-19 17:29:16 +00:00
|
|
|
on_time = genpd->on_time + delta;
|
|
|
|
do_div(on_time, NSEC_PER_MSEC);
|
|
|
|
seq_printf(s, "%llu ms\n", on_time);
|
2017-07-14 17:10:16 +00:00
|
|
|
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2018-12-15 08:45:26 +00:00
|
|
|
static int total_idle_time_show(struct seq_file *s, void *data)
|
2017-07-14 17:10:16 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = s->private;
|
2022-04-19 17:29:16 +00:00
|
|
|
u64 now, delta, total = 0;
|
2017-07-14 17:10:16 +00:00
|
|
|
unsigned int i;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
ret = genpd_lock_interruptible(genpd);
|
|
|
|
if (ret)
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
|
|
|
|
for (i = 0; i < genpd->state_count; i++) {
|
2022-04-19 17:29:16 +00:00
|
|
|
total += genpd->states[i].idle_time;
|
2017-07-14 17:10:16 +00:00
|
|
|
|
2022-04-19 17:29:16 +00:00
|
|
|
if (genpd->status == GENPD_STATE_OFF && genpd->state_idx == i) {
|
|
|
|
now = ktime_get_mono_fast_ns();
|
|
|
|
if (now > genpd->accounting_time) {
|
|
|
|
delta = now - genpd->accounting_time;
|
|
|
|
total += delta;
|
|
|
|
}
|
|
|
|
}
|
2017-07-14 17:10:16 +00:00
|
|
|
}
|
|
|
|
|
2022-04-19 17:29:16 +00:00
|
|
|
do_div(total, NSEC_PER_MSEC);
|
|
|
|
seq_printf(s, "%llu ms\n", total);
|
2017-07-14 17:10:16 +00:00
|
|
|
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2018-12-15 08:45:26 +00:00
|
|
|
static int devices_show(struct seq_file *s, void *data)
|
2017-07-14 17:10:16 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = s->private;
|
|
|
|
struct pm_domain_data *pm_data;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
ret = genpd_lock_interruptible(genpd);
|
|
|
|
if (ret)
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
|
2024-05-27 14:25:53 +00:00
|
|
|
list_for_each_entry(pm_data, &genpd->dev_list, list_node)
|
|
|
|
seq_printf(s, "%s\n", dev_name(pm_data->dev));
|
2017-07-14 17:10:16 +00:00
|
|
|
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2018-12-15 08:45:26 +00:00
|
|
|
static int perf_state_show(struct seq_file *s, void *data)
|
2018-05-30 09:45:17 +00:00
|
|
|
{
|
|
|
|
struct generic_pm_domain *genpd = s->private;
|
|
|
|
|
|
|
|
if (genpd_lock_interruptible(genpd))
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
|
|
|
|
seq_printf(s, "%u\n", genpd->performance_state);
|
|
|
|
|
|
|
|
genpd_unlock(genpd);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-12-15 08:45:26 +00:00
|
|
|
DEFINE_SHOW_ATTRIBUTE(summary);
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(status);
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(sub_domains);
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(idle_states);
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(active_time);
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(total_idle_time);
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(devices);
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(perf_state);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
|
2020-12-08 19:19:55 +00:00
|
|
|
static void genpd_debug_add(struct generic_pm_domain *genpd)
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
{
|
|
|
|
struct dentry *d;
|
2020-12-08 19:19:55 +00:00
|
|
|
|
|
|
|
if (!genpd_debugfs_dir)
|
|
|
|
return;
|
|
|
|
|
2024-10-30 12:55:10 +00:00
|
|
|
d = debugfs_create_dir(dev_name(&genpd->dev), genpd_debugfs_dir);
|
2020-12-08 19:19:55 +00:00
|
|
|
|
|
|
|
debugfs_create_file("current_state", 0444,
|
|
|
|
d, genpd, &status_fops);
|
|
|
|
debugfs_create_file("sub_domains", 0444,
|
|
|
|
d, genpd, &sub_domains_fops);
|
|
|
|
debugfs_create_file("idle_states", 0444,
|
|
|
|
d, genpd, &idle_states_fops);
|
|
|
|
debugfs_create_file("active_time", 0444,
|
|
|
|
d, genpd, &active_time_fops);
|
|
|
|
debugfs_create_file("total_idle_time", 0444,
|
|
|
|
d, genpd, &total_idle_time_fops);
|
|
|
|
debugfs_create_file("devices", 0444,
|
|
|
|
d, genpd, &devices_fops);
|
|
|
|
if (genpd->set_performance_state)
|
|
|
|
debugfs_create_file("perf_state", 0444,
|
|
|
|
d, genpd, &perf_state_fops);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int __init genpd_debug_init(void)
|
|
|
|
{
|
2017-07-14 17:10:16 +00:00
|
|
|
struct generic_pm_domain *genpd;
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
|
2017-10-06 07:02:06 +00:00
|
|
|
genpd_debugfs_dir = debugfs_create_dir("pm_genpd", NULL);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
|
2019-01-22 15:21:05 +00:00
|
|
|
debugfs_create_file("pm_genpd_summary", S_IRUGO, genpd_debugfs_dir,
|
|
|
|
NULL, &summary_fops);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
|
2020-12-08 19:19:55 +00:00
|
|
|
list_for_each_entry(genpd, &gpd_list, gpd_list_node)
|
|
|
|
genpd_debug_add(genpd);
|
2017-07-14 17:10:16 +00:00
|
|
|
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
return 0;
|
|
|
|
}
|
2017-10-06 07:02:06 +00:00
|
|
|
late_initcall(genpd_debug_init);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
|
2017-10-06 07:02:06 +00:00
|
|
|
static void __exit genpd_debug_exit(void)
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
{
|
2017-10-06 07:02:06 +00:00
|
|
|
debugfs_remove_recursive(genpd_debugfs_dir);
|
PM / Domains: add debugfs listing of struct generic_pm_domain-s
Add /sys/kernel/debug/pm_genpd/pm_genpd_summary file, which
lists power domains in the system, their statuses and attached devices,
resembling /sys/kernel/debug/clk/clk_summary.
Currently it is impossible to inspect (from userland) whether
a power domain is on or off. And, if it is on, which device blocks it
from powering down. This change allows developers working on
embedded devices power efficiency to list all necessary information
about generic power domains in one place.
The content of pm_genpd/pm_genpd_summary file is generated by iterating
over all generic power domain in the system, and, for each,
over registered devices and over the subdomains, if present.
Example output:
$ cat /sys/kernel/debug/pm_genpd/pm_genpd_summary
domain status slaves
/device runtime status
----------------------------------------------------------------------
a4su off
a3sg off
a3sm on
a3sp on
/devices/e6600000.pwm suspended
/devices/e6c50000.serial active
/devices/e6850000.sd suspended
/devices/e6bd0000.mmc active
a4s on a3sp, a3sm, a3sg
/devices/e6900000.irqpin unsupported
/devices/e6900004.irqpin unsupported
/devices/e6900008.irqpin unsupported
/devices/e690000c.irqpin unsupported
/devices/e9a00000.ethernet active
a3rv off
a4r off a3rv
/devices/fff20000.i2c suspended
a4lc off
c5 on a4lc, a4r, a4s, a4su
/devices/e6050000.pfc unsupported
/devices/e6138000.timer active
To enable this feature, compile the kernel with debugfs
and CONFIG_PM_ADVANCED_DEBUG enabled.
Signed-off-by: Maciej Matraszek <m.matraszek@samsung.com>
Tested-by: Geert Uytterhoeven <geert+renesas@glider.be>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2014-09-15 11:09:10 +00:00
|
|
|
}
|
2017-10-06 07:02:06 +00:00
|
|
|
__exitcall(genpd_debug_exit);
|
2016-08-11 10:40:05 +00:00
|
|
|
#endif /* CONFIG_DEBUG_FS */
|