staticvoid clk_pm_runtime_put(struct clk_core *core)
{ if (!core->rpm_enabled) return;
pm_runtime_put_sync(core->dev);
}
/** * clk_pm_runtime_get_all() - Runtime "get" all clk provider devices * * Call clk_pm_runtime_get() on all runtime PM enabled clks in the clk tree so * that disabling unused clks avoids a deadlock where a device is runtime PM * resuming/suspending and the runtime PM callback is trying to grab the * prepare_lock for something like clk_prepare_enable() while * clk_disable_unused_subtree() holds the prepare_lock and is trying to runtime * PM resume/suspend the device as well. * * Context: Acquires the 'clk_rpm_list_lock' and returns with the lock held on * success. Otherwise the lock is released on failure. * * Return: 0 on success, negative errno otherwise.
*/ staticint clk_pm_runtime_get_all(void)
{ int ret; struct clk_core *core, *failed;
/* * Grab the list lock to prevent any new clks from being registered * or unregistered until clk_pm_runtime_put_all().
*/
mutex_lock(&clk_rpm_list_lock);
/* * Runtime PM "get" all the devices that are needed for the clks * currently registered. Do this without holding the prepare_lock, to * avoid the deadlock.
*/
hlist_for_each_entry(core, &clk_rpm_list, rpm_node) {
ret = clk_pm_runtime_get(core); if (ret) {
failed = core;
pr_err("clk: Failed to runtime PM get '%s' for clk '%s'\n",
dev_name(failed->dev), failed->name); goto err;
}
}
return 0;
err:
hlist_for_each_entry(core, &clk_rpm_list, rpm_node) { if (core == failed) break;
/** * clk_pm_runtime_put_all() - Runtime "put" all clk provider devices * * Put the runtime PM references taken in clk_pm_runtime_get_all() and release * the 'clk_rpm_list_lock'.
*/ staticvoid clk_pm_runtime_put_all(void)
{ struct clk_core *core;
/* * On UP systems, spin_trylock_irqsave() always returns true, even if * we already hold the lock. So, in that case, we rely only on * reference counting.
*/ if (!IS_ENABLED(CONFIG_SMP) ||
!spin_trylock_irqsave(&enable_lock, flags)) { if (enable_owner == current) {
enable_refcnt++;
__acquire(enable_lock); if (!IS_ENABLED(CONFIG_SMP))
local_save_flags(flags); return flags;
}
spin_lock_irqsave(&enable_lock, flags);
}
WARN_ON_ONCE(enable_owner != NULL);
WARN_ON_ONCE(enable_refcnt != 0);
enable_owner = current;
enable_refcnt = 1; return flags;
}
staticbool clk_core_is_prepared(struct clk_core *core)
{ bool ret = false;
/* * .is_prepared is optional for clocks that can prepare * fall back to software usage counter if it is missing
*/ if (!core->ops->is_prepared) return core->prepare_count;
if (!clk_pm_runtime_get(core)) {
ret = core->ops->is_prepared(core->hw);
clk_pm_runtime_put(core);
}
return ret;
}
staticbool clk_core_is_enabled(struct clk_core *core)
{ bool ret = false;
/* * .is_enabled is only mandatory for clocks that gate * fall back to software usage counter if .is_enabled is missing
*/ if (!core->ops->is_enabled) return core->enable_count;
/* * Check if clock controller's device is runtime active before * calling .is_enabled callback. If not, assume that clock is * disabled, because we might be called from atomic context, from * which pm_runtime_get() is not allowed. * This function is called mainly from clk_disable_unused_subtree, * which ensures proper runtime pm activation of controller before * taking enable spinlock, but the below check is needed if one tries * to call it from other places.
*/ if (core->rpm_enabled) {
pm_runtime_get_noresume(core->dev); if (!pm_runtime_active(core->dev)) {
ret = false; goto done;
}
}
/* * This could be called with the enable lock held, or from atomic * context. If the parent isn't enabled already, we can't do * anything here. We can also assume this clock isn't enabled.
*/ if ((core->flags & CLK_OPS_PARENT_ENABLE) && core->parent) if (!clk_core_is_enabled(core->parent)) {
ret = false; goto done;
}
ret = core->ops->is_enabled(core->hw);
done: if (core->rpm_enabled)
pm_runtime_put(core->dev);
/* search the 'proper' clk tree first */
hlist_for_each_entry(root_clk, &clk_root_list, child_node) {
ret = __clk_lookup_subtree(name, root_clk); if (ret) return ret;
}
/* if not found, then search the orphan tree */
hlist_for_each_entry(root_clk, &clk_orphan_list, child_node) {
ret = __clk_lookup_subtree(name, root_clk); if (ret) return ret;
}
/** * clk_core_get - Find the clk_core parent of a clk * @core: clk to find parent of * @p_index: parent index to search for * * This is the preferred method for clk providers to find the parent of a * clk when that parent is external to the clk controller. The parent_names * array is indexed and treated as a local name matching a string in the device * node's 'clock-names' property or as the 'con_id' matching the device's * dev_name() in a clk_lookup. This allows clk providers to use their own * namespace instead of looking for a globally unique parent string. * * For example the following DT snippet would allow a clock registered by the * clock-controller@c001 that has a clk_init_data::parent_data array * with 'xtal' in the 'name' member to find the clock provided by the * clock-controller@f00abcd without needing to get the globally unique name of * the xtal clk. * * parent: clock-controller@f00abcd { * reg = <0xf00abcd 0xabcd>; * #clock-cells = <0>; * }; * * clock-controller@c001 { * reg = <0xc001 0xf00d>; * clocks = <&parent>; * clock-names = "xtal"; * #clock-cells = <1>; * }; * * Returns: -ENOENT when the provider can't be found or the clk doesn't * exist in the provider or the name can't be found in the DT node or * in a clkdev lookup. NULL when the provider knows about the clk but it * isn't provided on this system. * A valid clk_core pointer when the clk can be found in the provider.
*/ staticstruct clk_core *clk_core_get(struct clk_core *core, u8 p_index)
{ constchar *name = core->parents[p_index].fw_name; int index = core->parents[p_index].index; struct clk_hw *hw = ERR_PTR(-ENOENT); struct device *dev = core->dev; constchar *dev_id = dev ? dev_name(dev) : NULL; struct device_node *np = core->of_node; struct of_phandle_args clkspec;
if (np && (name || index >= 0) &&
!of_parse_clkspec(np, index, name, &clkspec)) {
hw = of_clk_get_hw_from_clkspec(&clkspec);
of_node_put(clkspec.np);
} elseif (name) { /* * If the DT search above couldn't find the provider fallback to * looking up via clkdev based clk_lookups.
*/
hw = clk_find_hw(dev_id, name);
}
/* * We have a direct reference but it isn't registered yet? * Orphan it and let clk_reparent() update the orphan status * when the parent is registered.
*/ if (!parent)
parent = ERR_PTR(-EPROBE_DEFER);
/* Only cache it if it's not an error */ if (!IS_ERR(parent))
entry->core = parent;
}
staticstruct clk_core *clk_core_get_parent_by_index(struct clk_core *core,
u8 index)
{ if (!core || index >= core->num_parents || !core->parents) return NULL;
if (!core->parents[index].core)
clk_core_fill_parent_index(core, index);
staticunsignedlong clk_core_get_rate_nolock(struct clk_core *core)
{ if (!core) return 0;
if (!core->num_parents || core->parent) return core->rate;
/* * Clk must have a parent because num_parents > 0 but the parent isn't * known yet. Best to return 0 as the rate of this clk until we can * properly recalc the rate based on the parent's rate.
*/ return 0;
}
ret = clk_core_round_rate_nolock(parent, &parent_req); if (ret) return ret;
trace_clk_rate_request_done(&parent_req);
best = parent_req.rate;
} elseif (parent) {
best = clk_core_get_rate_nolock(parent);
} else {
best = clk_core_get_rate_nolock(core);
}
req->best_parent_rate = best;
req->rate = best;
return 0;
}
int clk_mux_determine_rate_flags(struct clk_hw *hw, struct clk_rate_request *req, unsignedlong flags)
{ struct clk_core *core = hw->core, *parent, *best_parent = NULL; int i, num_parents, ret; unsignedlong best = 0;
/* if NO_REPARENT flag set, pass through to current parent */ if (core->flags & CLK_SET_RATE_NO_REPARENT) return clk_core_determine_rate_no_reparent(hw, req);
/* find the parent that can provide the fastest rate <= rate */
num_parents = core->num_parents; for (i = 0; i < num_parents; i++) { unsignedlong parent_rate;
parent = clk_core_get_parent_by_index(core, i); if (!parent) continue;
if (core->flags & CLK_SET_RATE_PARENT) { struct clk_rate_request parent_req;
/* * clk_hw_get_rate_range() - returns the clock rate range for a hw clk * @hw: the hw clk we want to get the range from * @min_rate: pointer to the variable that will hold the minimum * @max_rate: pointer to the variable that will hold the maximum * * Fills the @min_rate and @max_rate variables with the minimum and * maximum that clock can reach.
*/ void clk_hw_get_rate_range(struct clk_hw *hw, unsignedlong *min_rate, unsignedlong *max_rate)
{
clk_core_get_boundaries(hw->core, min_rate, max_rate);
}
EXPORT_SYMBOL_GPL(clk_hw_get_rate_range);
/* * __clk_mux_determine_rate - clk_ops::determine_rate implementation for a mux type clk * @hw: mux type clk to determine rate on * @req: rate request, also used to return preferred parent and frequencies * * Helper for finding best parent to provide a given frequency. This can be used * directly as a determine_rate callback (e.g. for a mux), or from a more * complex clock that may combine a mux with other operations. * * Returns: 0 on success, -EERROR value on error
*/ int __clk_mux_determine_rate(struct clk_hw *hw, struct clk_rate_request *req)
{ return clk_mux_determine_rate_flags(hw, req, 0);
}
EXPORT_SYMBOL_GPL(__clk_mux_determine_rate);
/* * clk_hw_determine_rate_no_reparent - clk_ops::determine_rate implementation for a clk that doesn't reparent * @hw: mux type clk to determine rate on * @req: rate request, also used to return preferred frequency * * Helper for finding best parent rate to provide a given frequency. * This can be used directly as a determine_rate callback (e.g. for a * mux), or from a more complex clock that may combine a mux with other * operations. * * Returns: 0 on success, -EERROR value on error
*/ int clk_hw_determine_rate_no_reparent(struct clk_hw *hw, struct clk_rate_request *req)
{ return clk_core_determine_rate_no_reparent(hw, req);
}
EXPORT_SYMBOL_GPL(clk_hw_determine_rate_no_reparent);
if (WARN(core->protect_count == 0, "%s already unprotected\n", core->name)) return;
if (--core->protect_count > 0) return;
clk_core_rate_unprotect(core->parent);
}
staticint clk_core_rate_nuke_protect(struct clk_core *core)
{ int ret;
lockdep_assert_held(&prepare_lock);
if (!core) return -EINVAL;
if (core->protect_count == 0) return 0;
ret = core->protect_count;
core->protect_count = 1;
clk_core_rate_unprotect(core);
return ret;
}
/** * clk_rate_exclusive_put - release exclusivity over clock rate control * @clk: the clk over which the exclusivity is released * * clk_rate_exclusive_put() completes a critical section during which a clock * consumer cannot tolerate any other consumer making any operation on the * clock which could result in a rate change or rate glitch. Exclusive clocks * cannot have their rate changed, either directly or indirectly due to changes * further up the parent chain of clocks. As a result, clocks up parent chain * also get under exclusive control of the calling consumer. * * If exlusivity is claimed more than once on clock, even by the same consumer, * the rate effectively gets locked as exclusivity can't be preempted. * * Calls to clk_rate_exclusive_put() must be balanced with calls to * clk_rate_exclusive_get(). Calls to this function may sleep, and do not return * error status.
*/ void clk_rate_exclusive_put(struct clk *clk)
{ if (!clk) return;
clk_prepare_lock();
/* * if there is something wrong with this consumer protect count, stop * here before messing with the provider
*/ if (WARN_ON(clk->exclusive_count <= 0)) goto out;
/** * clk_rate_exclusive_get - get exclusivity over the clk rate control * @clk: the clk over which the exclusity of rate control is requested * * clk_rate_exclusive_get() begins a critical section during which a clock * consumer cannot tolerate any other consumer making any operation on the * clock which could result in a rate change or rate glitch. Exclusive clocks * cannot have their rate changed, either directly or indirectly due to changes * further up the parent chain of clocks. As a result, clocks up parent chain * also get under exclusive control of the calling consumer. * * If exlusivity is claimed more than once on clock, even by the same consumer, * the rate effectively gets locked as exclusivity can't be preempted. * * Calls to clk_rate_exclusive_get() should be balanced with calls to * clk_rate_exclusive_put(). Calls to this function may sleep. * Returns 0 on success, -EERROR otherwise
*/ int clk_rate_exclusive_get(struct clk *clk)
{ if (!clk) return 0;
/** * clk_unprepare - undo preparation of a clock source * @clk: the clk being unprepared * * clk_unprepare may sleep, which differentiates it from clk_disable. In a * simple case, clk_unprepare can be used instead of clk_disable to gate a clk * if the operation may sleep. One example is a clk which is accessed over * I2c. In the complex case a clk gate operation may require a fast and a slow * part. It is this reason that clk_unprepare and clk_disable are not mutually * exclusive. In fact clk_disable must be called before clk_unprepare.
*/ void clk_unprepare(struct clk *clk)
{ if (IS_ERR_OR_NULL(clk)) return;
staticint clk_core_prepare(struct clk_core *core)
{ int ret = 0;
lockdep_assert_held(&prepare_lock);
if (!core) return 0;
if (core->prepare_count == 0) {
ret = clk_pm_runtime_get(core); if (ret) return ret;
ret = clk_core_prepare(core->parent); if (ret) goto runtime_put;
trace_clk_prepare(core);
if (core->ops->prepare)
ret = core->ops->prepare(core->hw);
trace_clk_prepare_complete(core);
if (ret) goto unprepare;
}
core->prepare_count++;
/* * CLK_SET_RATE_GATE is a special case of clock protection * Instead of a consumer claiming exclusive rate control, it is * actually the provider which prevents any consumer from making any * operation which could result in a rate change or rate glitch while * the clock is prepared.
*/ if (core->flags & CLK_SET_RATE_GATE)
clk_core_rate_protect(core);
staticint clk_core_prepare_lock(struct clk_core *core)
{ int ret;
clk_prepare_lock();
ret = clk_core_prepare(core);
clk_prepare_unlock();
return ret;
}
/** * clk_prepare - prepare a clock source * @clk: the clk being prepared * * clk_prepare may sleep, which differentiates it from clk_enable. In a simple * case, clk_prepare can be used instead of clk_enable to ungate a clk if the * operation may sleep. One example is a clk which is accessed over I2c. In * the complex case a clk ungate operation may require a fast and a slow part. * It is this reason that clk_prepare and clk_enable are not mutually * exclusive. In fact clk_prepare must be called before clk_enable. * Returns 0 on success, -EERROR otherwise.
*/ int clk_prepare(struct clk *clk)
{ if (!clk) return 0;
/** * clk_disable - gate a clock * @clk: the clk being gated * * clk_disable must not sleep, which differentiates it from clk_unprepare. In * a simple case, clk_disable can be used instead of clk_unprepare to gate a * clk if the operation is fast and will never sleep. One example is a * SoC-internal clk which is controlled via simple register writes. In the * complex case a clk gate operation may require a fast and a slow part. It is * this reason that clk_unprepare and clk_disable are not mutually exclusive. * In fact clk_disable must be called before clk_unprepare.
*/ void clk_disable(struct clk *clk)
{ if (IS_ERR_OR_NULL(clk)) return;
staticint clk_core_enable(struct clk_core *core)
{ int ret = 0;
lockdep_assert_held(&enable_lock);
if (!core) return 0;
if (WARN(core->prepare_count == 0, "Enabling unprepared %s\n", core->name)) return -ESHUTDOWN;
if (core->enable_count == 0) {
ret = clk_core_enable(core->parent);
if (ret) return ret;
trace_clk_enable(core);
if (core->ops->enable)
ret = core->ops->enable(core->hw);
trace_clk_enable_complete(core);
if (ret) {
clk_core_disable(core->parent); return ret;
}
}
core->enable_count++; return 0;
}
staticint clk_core_enable_lock(struct clk_core *core)
{ unsignedlong flags; int ret;
flags = clk_enable_lock();
ret = clk_core_enable(core);
clk_enable_unlock(flags);
return ret;
}
/** * clk_gate_restore_context - restore context for poweroff * @hw: the clk_hw pointer of clock whose state is to be restored * * The clock gate restore context function enables or disables * the gate clocks based on the enable_count. This is done in cases * where the clock context is lost and based on the enable_count * the clock either needs to be enabled/disabled. This * helps restore the state of gate clocks.
*/ void clk_gate_restore_context(struct clk_hw *hw)
{ struct clk_core *core = hw->core;
if (core->enable_count)
core->ops->enable(hw); else
core->ops->disable(hw);
}
EXPORT_SYMBOL_GPL(clk_gate_restore_context);
staticint clk_core_save_context(struct clk_core *core)
{ struct clk_core *child; int ret = 0;
hlist_for_each_entry(child, &core->children, child_node) {
ret = clk_core_save_context(child); if (ret < 0) return ret;
}
if (core->ops && core->ops->save_context)
ret = core->ops->save_context(core->hw);
/** * clk_save_context - save clock context for poweroff * * Saves the context of the clock register for powerstates in which the * contents of the registers will be lost. Occurs deep within the suspend * code. Returns 0 on success.
*/ int clk_save_context(void)
{ struct clk_core *clk; int ret;
hlist_for_each_entry(clk, &clk_root_list, child_node) {
ret = clk_core_save_context(clk); if (ret < 0) return ret;
}
hlist_for_each_entry(clk, &clk_orphan_list, child_node) {
ret = clk_core_save_context(clk); if (ret < 0) return ret;
}
return 0;
}
EXPORT_SYMBOL_GPL(clk_save_context);
/** * clk_restore_context - restore clock context after poweroff * * Restore the saved clock context upon resume. *
*/ void clk_restore_context(void)
{ struct clk_core *core;
/** * clk_enable - ungate a clock * @clk: the clk being ungated * * clk_enable must not sleep, which differentiates it from clk_prepare. In a * simple case, clk_enable can be used instead of clk_prepare to ungate a clk * if the operation will never sleep. One example is a SoC-internal clk which * is controlled via simple register writes. In the complex case a clk ungate * operation may require a fast and a slow part. It is this reason that * clk_enable and clk_prepare are not mutually exclusive. In fact clk_prepare * must be called before clk_enable. Returns 0 on success, -EERROR * otherwise.
*/ int clk_enable(struct clk *clk)
{ if (!clk) return 0;
/** * clk_is_enabled_when_prepared - indicate if preparing a clock also enables it. * @clk: clock source * * Returns true if clk_prepare() implicitly enables the clock, effectively * making clk_enable()/clk_disable() no-ops, false otherwise. * * This is of interest mainly to power management code where actually * disabling the clock also requires unpreparing it to have any material * effect. * * Regardless of the value returned here, the caller must always invoke * clk_enable() or clk_prepare_enable() and counterparts for usage counts * to be right.
*/ bool clk_is_enabled_when_prepared(struct clk *clk)
{ return clk && !(clk->core->ops->enable && clk->core->ops->disable);
}
EXPORT_SYMBOL_GPL(clk_is_enabled_when_prepared);
staticint clk_core_prepare_enable(struct clk_core *core)
{ int ret;
ret = clk_core_prepare_lock(core); if (ret) return ret;
ret = clk_core_enable_lock(core); if (ret)
clk_core_unprepare_lock(core);
if (core->flags & CLK_OPS_PARENT_ENABLE)
clk_core_prepare_enable(core->parent);
flags = clk_enable_lock();
if (core->enable_count) goto unlock_out;
if (core->flags & CLK_IGNORE_UNUSED) goto unlock_out;
/* * some gate clocks have special needs during the disable-unused * sequence. call .disable_unused if available, otherwise fall * back to .disable
*/ if (clk_core_is_enabled(core)) {
trace_clk_disable(core); if (core->ops->disable_unused)
core->ops->disable_unused(core->hw); elseif (core->ops->disable)
core->ops->disable(core->hw);
trace_clk_disable_complete(core);
}
unlock_out:
clk_enable_unlock(flags); if (core->flags & CLK_OPS_PARENT_ENABLE)
clk_core_disable_unprepare(core->parent);
}
staticint __init clk_disable_unused(void)
{ struct clk_core *core; int ret;
if (clk_ignore_unused) {
pr_warn("clk: Not disabling unused clocks\n"); return 0;
}
pr_info("clk: Disabling unused clocks\n");
ret = clk_pm_runtime_get_all(); if (ret) return ret; /* * Grab the prepare lock to keep the clk topology stable while iterating * over clks.
*/
clk_prepare_lock();
staticint clk_core_determine_round_nolock(struct clk_core *core, struct clk_rate_request *req)
{ long rate;
lockdep_assert_held(&prepare_lock);
if (!core) return 0;
/* * Some clock providers hand-craft their clk_rate_requests and * might not fill min_rate and max_rate. * * If it's the case, clamping the rate is equivalent to setting * the rate to 0 which is bad. Skip the clamping but complain so * that it gets fixed, hopefully.
*/ if (!req->min_rate && !req->max_rate)
pr_warn("%s: %s: clk_rate_request has initialized min or max rate.\n",
__func__, core->name); else
req->rate = clamp(req->rate, req->min_rate, req->max_rate);
/* * At this point, core protection will be disabled * - if the provider is not protected at all * - if the calling consumer is the only one which has exclusivity * over the provider
*/ if (clk_core_rate_is_protected(core)) {
req->rate = core->rate;
} elseif (core->ops->determine_rate) { return core->ops->determine_rate(core->hw, req);
} elseif (core->ops->round_rate) {
rate = core->ops->round_rate(core->hw, req->rate,
&req->best_parent_rate); if (rate < 0) return rate;
/** * clk_hw_init_rate_request - Initializes a clk_rate_request * @hw: the clk for which we want to submit a rate request * @req: the clk_rate_request structure we want to initialise * @rate: the rate which is to be requested * * Initializes a clk_rate_request structure to submit to * __clk_determine_rate() or similar functions.
*/ void clk_hw_init_rate_request(conststruct clk_hw *hw, struct clk_rate_request *req, unsignedlong rate)
{ if (WARN_ON(!hw || !req)) return;
/** * clk_hw_forward_rate_request - Forwards a clk_rate_request to a clock's parent * @hw: the original clock that got the rate request * @old_req: the original clk_rate_request structure we want to forward * @parent: the clk we want to forward @old_req to * @req: the clk_rate_request structure we want to initialise * @parent_rate: The rate which is to be requested to @parent * * Initializes a clk_rate_request structure to submit to a clock parent * in __clk_determine_rate() or similar functions.
*/ void clk_hw_forward_rate_request(conststruct clk_hw *hw, conststruct clk_rate_request *old_req, conststruct clk_hw *parent, struct clk_rate_request *req, unsignedlong parent_rate)
{ if (WARN_ON(!hw || !old_req || !parent || !req)) return;
/** * __clk_determine_rate - get the closest rate actually supported by a clock * @hw: determine the rate of this clock * @req: target rate request * * Useful for clk_ops such as .set_rate and .determine_rate.
*/ int __clk_determine_rate(struct clk_hw *hw, struct clk_rate_request *req)
{ if (!hw) {
req->rate = 0; return 0;
}
/** * clk_hw_round_rate() - round the given rate for a hw clk * @hw: the hw clk for which we are rounding a rate * @rate: the rate which is to be rounded * * Takes in a rate as input and rounds it to a rate that the clk can actually * use. * * Context: prepare_lock must be held. * For clk providers to call from within clk_ops such as .round_rate, * .determine_rate. * * Return: returns rounded rate of hw clk if clk supports round_rate operation * else returns the parent rate.
*/ unsignedlong clk_hw_round_rate(struct clk_hw *hw, unsignedlong rate)
{ int ret; struct clk_rate_request req;
clk_core_init_rate_req(hw->core, &req, rate);
trace_clk_rate_request_start(&req);
ret = clk_core_round_rate_nolock(hw->core, &req); if (ret) return 0;
/** * clk_round_rate - round the given rate for a clk * @clk: the clk for which we are rounding a rate * @rate: the rate which is to be rounded * * Takes in a rate as input and rounds it to a rate that the clk can actually * use which is then returned. If clk doesn't support round_rate operation * then the parent rate is returned.
*/ long clk_round_rate(struct clk *clk, unsignedlong rate)
{ struct clk_rate_request req; int ret;
if (!clk) return 0;
clk_prepare_lock();
if (clk->exclusive_count)
clk_core_rate_unprotect(clk->core);
clk_core_init_rate_req(clk->core, &req, rate);
trace_clk_rate_request_start(&req);
ret = clk_core_round_rate_nolock(clk->core, &req);
trace_clk_rate_request_done(&req);
if (clk->exclusive_count)
clk_core_rate_protect(clk->core);
/** * __clk_notify - call clk notifier chain * @core: clk that is changing rate * @msg: clk notifier type (see include/linux/clk.h) * @old_rate: old clk rate * @new_rate: new clk rate * * Triggers a notifier call chain on the clk rate-change notification * for 'clk'. Passes a pointer to the struct clk and the previous * and current rates to the notifier callback. Intended to be called by * internal clock code only. Returns NOTIFY_DONE from the last driver * called if all went well, or NOTIFY_STOP or NOTIFY_BAD immediately if * a driver returns that.
*/ staticint __clk_notify(struct clk_core *core, unsignedlong msg, unsignedlong old_rate, unsignedlong new_rate)
{ struct clk_notifier *cn; struct clk_notifier_data cnd; int ret = NOTIFY_DONE;
cnd.old_rate = old_rate;
cnd.new_rate = new_rate;
list_for_each_entry(cn, &clk_notifier_list, node) { if (cn->clk->core == core) {
cnd.clk = cn->clk;
ret = srcu_notifier_call_chain(&cn->notifier_head, msg,
&cnd); if (ret & NOTIFY_STOP_MASK) return ret;
}
}
return ret;
}
/** * __clk_recalc_accuracies * @core: first clk in the subtree * * Walks the subtree of clks starting with clk and recalculates accuracies as * it goes. Note that if a clk does not implement the .recalc_accuracy * callback then it is assumed that the clock will take on the accuracy of its * parent.
*/ staticvoid __clk_recalc_accuracies(struct clk_core *core)
{ unsignedlong parent_accuracy = 0; struct clk_core *child;
lockdep_assert_held(&prepare_lock);
if (core->parent)
parent_accuracy = core->parent->accuracy;
if (core->ops->recalc_accuracy)
core->accuracy = core->ops->recalc_accuracy(core->hw,
parent_accuracy); else
core->accuracy = parent_accuracy;
/** * clk_get_accuracy - return the accuracy of clk * @clk: the clk whose accuracy is being returned * * Simply returns the cached accuracy of the clk, unless * CLK_GET_ACCURACY_NOCACHE flag is set, which means a recalc_rate will be * issued. * If clk is NULL then returns 0.
*/ long clk_get_accuracy(struct clk *clk)
{ long accuracy;
/** * __clk_recalc_rates * @core: first clk in the subtree * @update_req: Whether req_rate should be updated with the new rate * @msg: notification type (see include/linux/clk.h) * * Walks the subtree of clks starting with clk and recalculates rates as it * goes. Note that if a clk does not implement the .recalc_rate callback then * it is assumed that the clock will take on the rate of its parent. * * clk_recalc_rates also propagates the POST_RATE_CHANGE notification, * if necessary.
*/ staticvoid __clk_recalc_rates(struct clk_core *core, bool update_req, unsignedlong msg)
{ unsignedlong old_rate; unsignedlong parent_rate = 0; struct clk_core *child;
lockdep_assert_held(&prepare_lock);
old_rate = core->rate;
if (core->parent)
parent_rate = core->parent->rate;
core->rate = clk_recalc(core, parent_rate); if (update_req)
core->req_rate = core->rate;
/* * ignore NOTIFY_STOP and NOTIFY_BAD return values for POST_RATE_CHANGE * & ABORT_RATE_CHANGE notifiers
*/ if (core->notifier_count && msg)
__clk_notify(core, msg, old_rate, core->rate);
/** * clk_get_rate - return the rate of clk * @clk: the clk whose rate is being returned * * Simply returns the cached rate of the clk, unless CLK_GET_RATE_NOCACHE flag * is set, which means a recalc_rate will be issued. Can be called regardless of * the clock enabledness. If clk is NULL, or if an error occurred, then returns * 0.
*/ unsignedlong clk_get_rate(struct clk *clk)
{ unsignedlong rate;
staticint clk_fetch_parent_index(struct clk_core *core, struct clk_core *parent)
{ int i;
if (!parent) return -EINVAL;
for (i = 0; i < core->num_parents; i++) { /* Found it first try! */ if (core->parents[i].core == parent) return i;
/* Something else is here, so keep looking */ if (core->parents[i].core) continue;
/* Maybe core hasn't been cached but the hw is all we know? */ if (core->parents[i].hw) { if (core->parents[i].hw == parent->hw) break;
/* Didn't match, but we're expecting a clk_hw */ continue;
}
/* Maybe it hasn't been cached (clk_set_parent() path) */ if (parent == clk_core_get(core, i)) break;
/* Fallback to comparing globally unique names */ if (core->parents[i].name &&
!strcmp(parent->name, core->parents[i].name)) break;
}
if (i == core->num_parents) return -EINVAL;
core->parents[i].core = parent; return i;
}
/** * clk_hw_get_parent_index - return the index of the parent clock * @hw: clk_hw associated with the clk being consumed * * Fetches and returns the index of parent clock. Returns -EINVAL if the given * clock does not have a current parent.
*/ int clk_hw_get_parent_index(struct clk_hw *hw)
{ struct clk_hw *parent = clk_hw_get_parent(hw);
/* * Update the orphan status of @core and all its children.
*/ staticvoid clk_core_update_orphan_status(struct clk_core *core, bool is_orphan)
{ struct clk_core *child;
/* * 1. enable parents for CLK_OPS_PARENT_ENABLE clock * * 2. Migrate prepare state between parents and prevent race with * clk_enable(). * * If the clock is not prepared, then a race with * clk_enable/disable() is impossible since we already have the * prepare lock (future calls to clk_enable() need to be preceded by * a clk_prepare()). * * If the clock is prepared, migrate the prepared state to the new * parent and also protect against a race with clk_enable() by * forcing the clock and the new parent on. This ensures that all * future calls to clk_enable() are practically NOPs with respect to * hardware and software states. * * See also: Comment for clk_set_parent() below.
*/
/* enable old_parent & parent if CLK_OPS_PARENT_ENABLE is set */ if (core->flags & CLK_OPS_PARENT_ENABLE) {
clk_core_prepare_enable(old_parent);
clk_core_prepare_enable(parent);
}
/* migrate prepare count if > 0 */ if (core->prepare_count) {
clk_core_prepare_enable(parent);
clk_core_enable_lock(core);
}
/* update the clk tree topology */
flags = clk_enable_lock();
clk_reparent(core, parent);
clk_enable_unlock(flags);
return old_parent;
}
staticvoid __clk_set_parent_after(struct clk_core *core, struct clk_core *parent, struct clk_core *old_parent)
{ /* * Finish the migration of prepare state and undo the changes done * for preventing a race with clk_enable().
*/ if (core->prepare_count) {
clk_core_disable_lock(core);
clk_core_disable_unprepare(old_parent);
}
/* re-balance ref counting if CLK_OPS_PARENT_ENABLE is set */ if (core->flags & CLK_OPS_PARENT_ENABLE) {
clk_core_disable_unprepare(parent);
clk_core_disable_unprepare(old_parent);
}
}
staticint __clk_set_parent(struct clk_core *core, struct clk_core *parent,
u8 p_index)
{ unsignedlong flags; int ret = 0; struct clk_core *old_parent;
/* change clock input source */ if (parent && core->ops->set_parent)
ret = core->ops->set_parent(core->hw, p_index);
trace_clk_set_parent_complete(core, parent);
if (ret) {
flags = clk_enable_lock();
clk_reparent(core, old_parent);
clk_enable_unlock(flags);
__clk_set_parent_after(core, old_parent, parent);
return ret;
}
__clk_set_parent_after(core, parent, old_parent);
return 0;
}
/** * __clk_speculate_rates * @core: first clk in the subtree * @parent_rate: the "future" rate of clk's parent * * Walks the subtree of clks starting with clk, speculating rates as it * goes and firing off PRE_RATE_CHANGE notifications as necessary. * * Unlike clk_recalc_rates, clk_speculate_rates exists only for sending * pre-rate change notifications and returns early if no clks in the * subtree have subscribed to the notifications. Note that if a clk does not * implement the .recalc_rate callback then it is assumed that the clock will * take on the rate of its parent.
*/ staticint __clk_speculate_rates(struct clk_core *core, unsignedlong parent_rate)
{ struct clk_core *child; unsignedlong new_rate; int ret = NOTIFY_DONE;
lockdep_assert_held(&prepare_lock);
new_rate = clk_recalc(core, parent_rate);
/* abort rate change if a driver returns NOTIFY_BAD or NOTIFY_STOP */ if (core->notifier_count)
ret = __clk_notify(core, PRE_RATE_CHANGE, core->rate, new_rate);
if (ret & NOTIFY_STOP_MASK) {
pr_debug("%s: clk notifier callback for clock %s aborted with error %d\n",
__func__, core->name, ret); goto out;
}
hlist_for_each_entry(child, &core->children, child_node) {
ret = __clk_speculate_rates(child, new_rate); if (ret & NOTIFY_STOP_MASK) break;
}
/* some clocks must be gated to change parent */ if (parent != old_parent &&
(core->flags & CLK_SET_PARENT_GATE) && core->prepare_count) {
pr_debug("%s: %s not gated but wants to reparent\n",
__func__, core->name); return NULL;
}
/* try finding the new parent index */ if (parent && core->num_parents > 1) {
p_index = clk_fetch_parent_index(core, parent); if (p_index < 0) {
pr_debug("%s: clk %s can not be parent of clk %s\n",
__func__, parent->name, core->name); return NULL;
}
}
if ((core->flags & CLK_SET_RATE_PARENT) && parent &&
best_parent_rate != parent->rate)
top = clk_calc_new_rates(parent, best_parent_rate);
/* * Notify about rate changes in a subtree. Always walk down the whole tree * so that in case of an error we can walk down the whole tree again and * abort the change.
*/ staticstruct clk_core *clk_propagate_rate_change(struct clk_core *core, unsignedlong event)
{ struct clk_core *child, *tmp_clk, *fail_clk = NULL; int ret = NOTIFY_DONE;
if (core->rate == core->new_rate) return NULL;
if (core->notifier_count) {
ret = __clk_notify(core, event, core->rate, core->new_rate); if (ret & NOTIFY_STOP_MASK)
fail_clk = core;
}
hlist_for_each_entry(child, &core->children, child_node) { /* Skip children who will be reparented to another clock */ if (child->new_parent && child->new_parent != core) continue;
tmp_clk = clk_propagate_rate_change(child, event); if (tmp_clk)
fail_clk = tmp_clk;
}
/* handle the new child who might not be in core->children yet */ if (core->new_child) {
tmp_clk = clk_propagate_rate_change(core->new_child, event); if (tmp_clk)
fail_clk = tmp_clk;
}
return fail_clk;
}
/* * walk down a subtree and set the new rates notifying the rate * change on the way
*/ staticvoid clk_change_rate(struct clk_core *core)
{ struct clk_core *child; struct hlist_node *tmp; unsignedlong old_rate; unsignedlong best_parent_rate = 0; bool skip_set_rate = false; struct clk_core *old_parent; struct clk_core *parent = NULL;
if (core->flags & CLK_SET_RATE_UNGATE) {
clk_core_disable_lock(core);
clk_core_unprepare(core);
}
if (core->flags & CLK_OPS_PARENT_ENABLE)
clk_core_disable_unprepare(parent);
if (core->notifier_count && old_rate != core->rate)
__clk_notify(core, POST_RATE_CHANGE, old_rate, core->rate);
if (core->flags & CLK_RECALC_NEW_RATES)
(void)clk_calc_new_rates(core, core->new_rate);
/* * Use safe iteration, as change_rate can actually swap parents * for certain clock types.
*/
hlist_for_each_entry_safe(child, tmp, &core->children, child_node) { /* Skip children who will be reparented to another clock */ if (child->new_parent && child->new_parent != core) continue;
clk_change_rate(child);
}
/* handle the new child who might not be in core->children yet */ if (core->new_child)
clk_change_rate(core->new_child);
/* bail early if nothing to do */ if (rate == clk_core_get_rate_nolock(core)) return 0;
/* fail on a direct rate set of a protected provider */ if (clk_core_rate_is_protected(core)) return -EBUSY;
/* calculate new rates and get the topmost changed clock */
top = clk_calc_new_rates(core, req_rate); if (!top) return -EINVAL;
ret = clk_pm_runtime_get(core); if (ret) return ret;
/* notify that we are about to change rates */
fail_clk = clk_propagate_rate_change(top, PRE_RATE_CHANGE); if (fail_clk) {
pr_debug("%s: failed to set %s rate\n", __func__,
fail_clk->name);
clk_propagate_rate_change(top, ABORT_RATE_CHANGE);
ret = -EBUSY; goto err;
}
/** * clk_set_rate - specify a new rate for clk * @clk: the clk whose rate is being changed * @rate: the new rate for clk * * In the simplest case clk_set_rate will only adjust the rate of clk. * * Setting the CLK_SET_RATE_PARENT flag allows the rate change operation to * propagate up to clk's parent; whether or not this happens depends on the * outcome of clk's .round_rate implementation. If *parent_rate is unchanged * after calling .round_rate then upstream parent propagation is ignored. If * *parent_rate comes back with a new rate for clk's parent then we propagate * up to clk's parent and set its rate. Upward propagation will continue * until either a clk does not support the CLK_SET_RATE_PARENT flag or * .round_rate stops requesting changes to clk's parent_rate. * * Rate changes are accomplished via tree traversal that also recalculates the * rates for the clocks and fires off POST_RATE_CHANGE notifiers.
--> --------------------
--> maximum size reached
--> --------------------
Messung V0.5
¤ Dauer der Verarbeitung: 0.33 Sekunden
(vorverarbeitet)
¤
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.