Orange Pi5 kernel

Deprecated Linux kernel 5.10.110 for OrangePi 5/5B/5+ boards

3 Commits   0 Branches   0 Tags   |
// SPDX-License-Identifier: GPL-2.0
/*
* Copyright (C) 2020 Intel Corporation.
* Zhu YiXin <yixin.zhu@intel.com>
* Rahul Tanwar <rahul.tanwar@intel.com>
*/
#include <linux/clk-provider.h>
#include <linux/device.h>
#include <linux/of.h>
#include "clk-cgu.h"
#define GATE_HW_REG_STAT(reg) ((reg) + 0x0)
#define GATE_HW_REG_EN(reg) ((reg) + 0x4)
#define GATE_HW_REG_DIS(reg) ((reg) + 0x8)
#define MAX_DDIV_REG 8
#define MAX_DIVIDER_VAL 64
#define to_lgm_clk_mux(_hw) container_of(_hw, struct lgm_clk_mux, hw)
#define to_lgm_clk_divider(_hw) container_of(_hw, struct lgm_clk_divider, hw)
#define to_lgm_clk_gate(_hw) container_of(_hw, struct lgm_clk_gate, hw)
#define to_lgm_clk_ddiv(_hw) container_of(_hw, struct lgm_clk_ddiv, hw)
static struct clk_hw *lgm_clk_register_fixed(struct lgm_clk_provider *ctx,
<------><------><------><------><------> const struct lgm_clk_branch *list)
{
<------>unsigned long flags;
<------>if (list->div_flags & CLOCK_FLAG_VAL_INIT) {
<------><------>spin_lock_irqsave(&ctx->lock, flags);
<------><------>lgm_set_clk_val(ctx->membase, list->div_off, list->div_shift,
<------><------><------><------>list->div_width, list->div_val);
<------><------>spin_unlock_irqrestore(&ctx->lock, flags);
<------>}
<------>return clk_hw_register_fixed_rate(NULL, list->name,
<------><------><------><------><------> list->parent_data[0].name,
<------><------><------><------><------> list->flags, list->mux_flags);
}
static u8 lgm_clk_mux_get_parent(struct clk_hw *hw)
{
<------>struct lgm_clk_mux *mux = to_lgm_clk_mux(hw);
<------>unsigned long flags;
<------>u32 val;
<------>spin_lock_irqsave(&mux->lock, flags);
<------>if (mux->flags & MUX_CLK_SW)
<------><------>val = mux->reg;
<------>else
<------><------>val = lgm_get_clk_val(mux->membase, mux->reg, mux->shift,
<------><------><------><------> mux->width);
<------>spin_unlock_irqrestore(&mux->lock, flags);
<------>return clk_mux_val_to_index(hw, NULL, mux->flags, val);
}
static int lgm_clk_mux_set_parent(struct clk_hw *hw, u8 index)
{
<------>struct lgm_clk_mux *mux = to_lgm_clk_mux(hw);
<------>unsigned long flags;
<------>u32 val;
<------>val = clk_mux_index_to_val(NULL, mux->flags, index);
<------>spin_lock_irqsave(&mux->lock, flags);
<------>if (mux->flags & MUX_CLK_SW)
<------><------>mux->reg = val;
<------>else
<------><------>lgm_set_clk_val(mux->membase, mux->reg, mux->shift,
<------><------><------><------>mux->width, val);
<------>spin_unlock_irqrestore(&mux->lock, flags);
<------>return 0;
}
static int lgm_clk_mux_determine_rate(struct clk_hw *hw,
<------><------><------><------> struct clk_rate_request *req)
{
<------>struct lgm_clk_mux *mux = to_lgm_clk_mux(hw);
<------>return clk_mux_determine_rate_flags(hw, req, mux->flags);
}
static const struct clk_ops lgm_clk_mux_ops = {
<------>.get_parent = lgm_clk_mux_get_parent,
<------>.set_parent = lgm_clk_mux_set_parent,
<------>.determine_rate = lgm_clk_mux_determine_rate,
};
static struct clk_hw *
lgm_clk_register_mux(struct lgm_clk_provider *ctx,
<------><------> const struct lgm_clk_branch *list)
{
<------>unsigned long flags, cflags = list->mux_flags;
<------>struct device *dev = ctx->dev;
<------>u8 shift = list->mux_shift;
<------>u8 width = list->mux_width;
<------>struct clk_init_data init = {};
<------>struct lgm_clk_mux *mux;
<------>u32 reg = list->mux_off;
<------>struct clk_hw *hw;
<------>int ret;
<------>mux = devm_kzalloc(dev, sizeof(*mux), GFP_KERNEL);
<------>if (!mux)
<------><------>return ERR_PTR(-ENOMEM);
<------>init.name = list->name;
<------>init.ops = &lgm_clk_mux_ops;
<------>init.flags = list->flags;
<------>init.parent_data = list->parent_data;
<------>init.num_parents = list->num_parents;
<------>mux->membase = ctx->membase;
<------>mux->lock = ctx->lock;
<------>mux->reg = reg;
<------>mux->shift = shift;
<------>mux->width = width;
<------>mux->flags = cflags;
<------>mux->hw.init = &init;
<------>hw = &mux->hw;
<------>ret = devm_clk_hw_register(dev, hw);
<------>if (ret)
<------><------>return ERR_PTR(ret);
<------>if (cflags & CLOCK_FLAG_VAL_INIT) {
<------><------>spin_lock_irqsave(&mux->lock, flags);
<------><------>lgm_set_clk_val(mux->membase, reg, shift, width, list->mux_val);
<------><------>spin_unlock_irqrestore(&mux->lock, flags);
<------>}
<------>return hw;
}
static unsigned long
lgm_clk_divider_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
{
<------>struct lgm_clk_divider *divider = to_lgm_clk_divider(hw);
<------>unsigned long flags;
<------>unsigned int val;
<------>spin_lock_irqsave(&divider->lock, flags);
<------>val = lgm_get_clk_val(divider->membase, divider->reg,
<------><------><------> divider->shift, divider->width);
<------>spin_unlock_irqrestore(&divider->lock, flags);
<------>return divider_recalc_rate(hw, parent_rate, val, divider->table,
<------><------><------><------> divider->flags, divider->width);
}
static long
lgm_clk_divider_round_rate(struct clk_hw *hw, unsigned long rate,
<------><------><------> unsigned long *prate)
{
<------>struct lgm_clk_divider *divider = to_lgm_clk_divider(hw);
<------>return divider_round_rate(hw, rate, prate, divider->table,
<------><------><------><------> divider->width, divider->flags);
}
static int
lgm_clk_divider_set_rate(struct clk_hw *hw, unsigned long rate,
<------><------><------> unsigned long prate)
{
<------>struct lgm_clk_divider *divider = to_lgm_clk_divider(hw);
<------>unsigned long flags;
<------>int value;
<------>value = divider_get_val(rate, prate, divider->table,
<------><------><------><------>divider->width, divider->flags);
<------>if (value < 0)
<------><------>return value;
<------>spin_lock_irqsave(&divider->lock, flags);
<------>lgm_set_clk_val(divider->membase, divider->reg,
<------><------><------>divider->shift, divider->width, value);
<------>spin_unlock_irqrestore(&divider->lock, flags);
<------>return 0;
}
static int lgm_clk_divider_enable_disable(struct clk_hw *hw, int enable)
{
<------>struct lgm_clk_divider *div = to_lgm_clk_divider(hw);
<------>unsigned long flags;
<------>spin_lock_irqsave(&div->lock, flags);
<------>lgm_set_clk_val(div->membase, div->reg, div->shift_gate,
<------><------><------>div->width_gate, enable);
<------>spin_unlock_irqrestore(&div->lock, flags);
<------>return 0;
}
static int lgm_clk_divider_enable(struct clk_hw *hw)
{
<------>return lgm_clk_divider_enable_disable(hw, 1);
}
static void lgm_clk_divider_disable(struct clk_hw *hw)
{
<------>lgm_clk_divider_enable_disable(hw, 0);
}
static const struct clk_ops lgm_clk_divider_ops = {
<------>.recalc_rate = lgm_clk_divider_recalc_rate,
<------>.round_rate = lgm_clk_divider_round_rate,
<------>.set_rate = lgm_clk_divider_set_rate,
<------>.enable = lgm_clk_divider_enable,
<------>.disable = lgm_clk_divider_disable,
};
static struct clk_hw *
lgm_clk_register_divider(struct lgm_clk_provider *ctx,
<------><------><------> const struct lgm_clk_branch *list)
{
<------>unsigned long flags, cflags = list->div_flags;
<------>struct device *dev = ctx->dev;
<------>struct lgm_clk_divider *div;
<------>struct clk_init_data init = {};
<------>u8 shift = list->div_shift;
<------>u8 width = list->div_width;
<------>u8 shift_gate = list->div_shift_gate;
<------>u8 width_gate = list->div_width_gate;
<------>u32 reg = list->div_off;
<------>struct clk_hw *hw;
<------>int ret;
<------>div = devm_kzalloc(dev, sizeof(*div), GFP_KERNEL);
<------>if (!div)
<------><------>return ERR_PTR(-ENOMEM);
<------>init.name = list->name;
<------>init.ops = &lgm_clk_divider_ops;
<------>init.flags = list->flags;
<------>init.parent_data = list->parent_data;
<------>init.num_parents = 1;
<------>div->membase = ctx->membase;
<------>div->lock = ctx->lock;
<------>div->reg = reg;
<------>div->shift = shift;
<------>div->width = width;
<------>div->shift_gate = shift_gate;
<------>div->width_gate = width_gate;
<------>div->flags = cflags;
<------>div->table = list->div_table;
<------>div->hw.init = &init;
<------>hw = &div->hw;
<------>ret = devm_clk_hw_register(dev, hw);
<------>if (ret)
<------><------>return ERR_PTR(ret);
<------>if (cflags & CLOCK_FLAG_VAL_INIT) {
<------><------>spin_lock_irqsave(&div->lock, flags);
<------><------>lgm_set_clk_val(div->membase, reg, shift, width, list->div_val);
<------><------>spin_unlock_irqrestore(&div->lock, flags);
<------>}
<------>return hw;
}
static struct clk_hw *
lgm_clk_register_fixed_factor(struct lgm_clk_provider *ctx,
<------><------><------> const struct lgm_clk_branch *list)
{
<------>unsigned long flags;
<------>struct clk_hw *hw;
<------>hw = clk_hw_register_fixed_factor(ctx->dev, list->name,
<------><------><------><------><------> list->parent_data[0].name, list->flags,
<------><------><------><------><------> list->mult, list->div);
<------>if (IS_ERR(hw))
<------><------>return ERR_CAST(hw);
<------>if (list->div_flags & CLOCK_FLAG_VAL_INIT) {
<------><------>spin_lock_irqsave(&ctx->lock, flags);
<------><------>lgm_set_clk_val(ctx->membase, list->div_off, list->div_shift,
<------><------><------><------>list->div_width, list->div_val);
<------><------>spin_unlock_irqrestore(&ctx->lock, flags);
<------>}
<------>return hw;
}
static int lgm_clk_gate_enable(struct clk_hw *hw)
{
<------>struct lgm_clk_gate *gate = to_lgm_clk_gate(hw);
<------>unsigned long flags;
<------>unsigned int reg;
<------>spin_lock_irqsave(&gate->lock, flags);
<------>reg = GATE_HW_REG_EN(gate->reg);
<------>lgm_set_clk_val(gate->membase, reg, gate->shift, 1, 1);
<------>spin_unlock_irqrestore(&gate->lock, flags);
<------>return 0;
}
static void lgm_clk_gate_disable(struct clk_hw *hw)
{
<------>struct lgm_clk_gate *gate = to_lgm_clk_gate(hw);
<------>unsigned long flags;
<------>unsigned int reg;
<------>spin_lock_irqsave(&gate->lock, flags);
<------>reg = GATE_HW_REG_DIS(gate->reg);
<------>lgm_set_clk_val(gate->membase, reg, gate->shift, 1, 1);
<------>spin_unlock_irqrestore(&gate->lock, flags);
}
static int lgm_clk_gate_is_enabled(struct clk_hw *hw)
{
<------>struct lgm_clk_gate *gate = to_lgm_clk_gate(hw);
<------>unsigned int reg, ret;
<------>unsigned long flags;
<------>spin_lock_irqsave(&gate->lock, flags);
<------>reg = GATE_HW_REG_STAT(gate->reg);
<------>ret = lgm_get_clk_val(gate->membase, reg, gate->shift, 1);
<------>spin_unlock_irqrestore(&gate->lock, flags);
<------>return ret;
}
static const struct clk_ops lgm_clk_gate_ops = {
<------>.enable = lgm_clk_gate_enable,
<------>.disable = lgm_clk_gate_disable,
<------>.is_enabled = lgm_clk_gate_is_enabled,
};
static struct clk_hw *
lgm_clk_register_gate(struct lgm_clk_provider *ctx,
<------><------> const struct lgm_clk_branch *list)
{
<------>unsigned long flags, cflags = list->gate_flags;
<------>const char *pname = list->parent_data[0].name;
<------>struct device *dev = ctx->dev;
<------>u8 shift = list->gate_shift;
<------>struct clk_init_data init = {};
<------>struct lgm_clk_gate *gate;
<------>u32 reg = list->gate_off;
<------>struct clk_hw *hw;
<------>int ret;
<------>gate = devm_kzalloc(dev, sizeof(*gate), GFP_KERNEL);
<------>if (!gate)
<------><------>return ERR_PTR(-ENOMEM);
<------>init.name = list->name;
<------>init.ops = &lgm_clk_gate_ops;
<------>init.flags = list->flags;
<------>init.parent_names = pname ? &pname : NULL;
<------>init.num_parents = pname ? 1 : 0;
<------>gate->membase = ctx->membase;
<------>gate->lock = ctx->lock;
<------>gate->reg = reg;
<------>gate->shift = shift;
<------>gate->flags = cflags;
<------>gate->hw.init = &init;
<------>hw = &gate->hw;
<------>ret = devm_clk_hw_register(dev, hw);
<------>if (ret)
<------><------>return ERR_PTR(ret);
<------>if (cflags & CLOCK_FLAG_VAL_INIT) {
<------><------>spin_lock_irqsave(&gate->lock, flags);
<------><------>lgm_set_clk_val(gate->membase, reg, shift, 1, list->gate_val);
<------><------>spin_unlock_irqrestore(&gate->lock, flags);
<------>}
<------>return hw;
}
int lgm_clk_register_branches(struct lgm_clk_provider *ctx,
<------><------><------> const struct lgm_clk_branch *list,
<------><------><------> unsigned int nr_clk)
{
<------>struct clk_hw *hw;
<------>unsigned int idx;
<------>for (idx = 0; idx < nr_clk; idx++, list++) {
<------><------>switch (list->type) {
<------><------>case CLK_TYPE_FIXED:
<------><------><------>hw = lgm_clk_register_fixed(ctx, list);
<------><------><------>break;
<------><------>case CLK_TYPE_MUX:
<------><------><------>hw = lgm_clk_register_mux(ctx, list);
<------><------><------>break;
<------><------>case CLK_TYPE_DIVIDER:
<------><------><------>hw = lgm_clk_register_divider(ctx, list);
<------><------><------>break;
<------><------>case CLK_TYPE_FIXED_FACTOR:
<------><------><------>hw = lgm_clk_register_fixed_factor(ctx, list);
<------><------><------>break;
<------><------>case CLK_TYPE_GATE:
<------><------><------>hw = lgm_clk_register_gate(ctx, list);
<------><------><------>break;
<------><------>default:
<------><------><------>dev_err(ctx->dev, "invalid clk type\n");
<------><------><------>return -EINVAL;
<------><------>}
<------><------>if (IS_ERR(hw)) {
<------><------><------>dev_err(ctx->dev,
<------><------><------><------>"register clk: %s, type: %u failed!\n",
<------><------><------><------>list->name, list->type);
<------><------><------>return -EIO;
<------><------>}
<------><------>ctx->clk_data.hws[list->id] = hw;
<------>}
<------>return 0;
}
static unsigned long
lgm_clk_ddiv_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
{
<------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
<------>unsigned int div0, div1, exdiv;
<------>u64 prate;
<------>div0 = lgm_get_clk_val(ddiv->membase, ddiv->reg,
<------><------><------> ddiv->shift0, ddiv->width0) + 1;
<------>div1 = lgm_get_clk_val(ddiv->membase, ddiv->reg,
<------><------><------> ddiv->shift1, ddiv->width1) + 1;
<------>exdiv = lgm_get_clk_val(ddiv->membase, ddiv->reg,
<------><------><------><------>ddiv->shift2, ddiv->width2);
<------>prate = (u64)parent_rate;
<------>do_div(prate, div0);
<------>do_div(prate, div1);
<------>if (exdiv) {
<------><------>do_div(prate, ddiv->div);
<------><------>prate *= ddiv->mult;
<------>}
<------>return prate;
}
static int lgm_clk_ddiv_enable(struct clk_hw *hw)
{
<------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
<------>unsigned long flags;
<------>spin_lock_irqsave(&ddiv->lock, flags);
<------>lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift_gate,
<------><------><------>ddiv->width_gate, 1);
<------>spin_unlock_irqrestore(&ddiv->lock, flags);
<------>return 0;
}
static void lgm_clk_ddiv_disable(struct clk_hw *hw)
{
<------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
<------>unsigned long flags;
<------>spin_lock_irqsave(&ddiv->lock, flags);
<------>lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift_gate,
<------><------><------>ddiv->width_gate, 0);
<------>spin_unlock_irqrestore(&ddiv->lock, flags);
}
static int
lgm_clk_get_ddiv_val(u32 div, u32 *ddiv1, u32 *ddiv2)
{
<------>u32 idx, temp;
<------>*ddiv1 = 1;
<------>*ddiv2 = 1;
<------>if (div > MAX_DIVIDER_VAL)
<------><------>div = MAX_DIVIDER_VAL;
<------>if (div > 1) {
<------><------>for (idx = 2; idx <= MAX_DDIV_REG; idx++) {
<------><------><------>temp = DIV_ROUND_UP_ULL((u64)div, idx);
<------><------><------>if (div % idx == 0 && temp <= MAX_DDIV_REG)
<------><------><------><------>break;
<------><------>}
<------><------>if (idx > MAX_DDIV_REG)
<------><------><------>return -EINVAL;
<------><------>*ddiv1 = temp;
<------><------>*ddiv2 = idx;
<------>}
<------>return 0;
}
static int
lgm_clk_ddiv_set_rate(struct clk_hw *hw, unsigned long rate,
<------><------> unsigned long prate)
{
<------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
<------>u32 div, ddiv1, ddiv2;
<------>unsigned long flags;
<------>div = DIV_ROUND_CLOSEST_ULL((u64)prate, rate);
<------>spin_lock_irqsave(&ddiv->lock, flags);
<------>if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) {
<------><------>div = DIV_ROUND_CLOSEST_ULL((u64)div, 5);
<------><------>div = div * 2;
<------>}
<------>if (div <= 0) {
<------><------>spin_unlock_irqrestore(&ddiv->lock, flags);
<------><------>return -EINVAL;
<------>}
<------>if (lgm_clk_get_ddiv_val(div, &ddiv1, &ddiv2)) {
<------><------>spin_unlock_irqrestore(&ddiv->lock, flags);
<------><------>return -EINVAL;
<------>}
<------>lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift0, ddiv->width0,
<------><------><------>ddiv1 - 1);
<------>lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift1, ddiv->width1,
<------><------><------>ddiv2 - 1);
<------>spin_unlock_irqrestore(&ddiv->lock, flags);
<------>return 0;
}
static long
lgm_clk_ddiv_round_rate(struct clk_hw *hw, unsigned long rate,
<------><------><------>unsigned long *prate)
{
<------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
<------>u32 div, ddiv1, ddiv2;
<------>unsigned long flags;
<------>u64 rate64;
<------>div = DIV_ROUND_CLOSEST_ULL((u64)*prate, rate);
<------>/* if predivide bit is enabled, modify div by factor of 2.5 */
<------>spin_lock_irqsave(&ddiv->lock, flags);
<------>if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) {
<------><------>div = div * 2;
<------><------>div = DIV_ROUND_CLOSEST_ULL((u64)div, 5);
<------>}
<------>spin_unlock_irqrestore(&ddiv->lock, flags);
<------>if (div <= 0)
<------><------>return *prate;
<------>if (lgm_clk_get_ddiv_val(div, &ddiv1, &ddiv2) != 0)
<------><------>if (lgm_clk_get_ddiv_val(div + 1, &ddiv1, &ddiv2) != 0)
<------><------><------>return -EINVAL;
<------>rate64 = *prate;
<------>do_div(rate64, ddiv1);
<------>do_div(rate64, ddiv2);
<------>/* if predivide bit is enabled, modify rounded rate by factor of 2.5 */
<------>spin_lock_irqsave(&ddiv->lock, flags);
<------>if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) {
<------><------>rate64 = rate64 * 2;
<------><------>rate64 = DIV_ROUND_CLOSEST_ULL(rate64, 5);
<------>}
<------>spin_unlock_irqrestore(&ddiv->lock, flags);
<------>return rate64;
}
static const struct clk_ops lgm_clk_ddiv_ops = {
<------>.recalc_rate = lgm_clk_ddiv_recalc_rate,
<------>.enable = lgm_clk_ddiv_enable,
<------>.disable = lgm_clk_ddiv_disable,
<------>.set_rate = lgm_clk_ddiv_set_rate,
<------>.round_rate = lgm_clk_ddiv_round_rate,
};
int lgm_clk_register_ddiv(struct lgm_clk_provider *ctx,
<------><------><------> const struct lgm_clk_ddiv_data *list,
<------><------><------> unsigned int nr_clk)
{
<------>struct device *dev = ctx->dev;
<------>struct clk_hw *hw;
<------>unsigned int idx;
<------>int ret;
<------>for (idx = 0; idx < nr_clk; idx++, list++) {
<------><------>struct clk_init_data init = {};
<------><------>struct lgm_clk_ddiv *ddiv;
<------><------>ddiv = devm_kzalloc(dev, sizeof(*ddiv), GFP_KERNEL);
<------><------>if (!ddiv)
<------><------><------>return -ENOMEM;
<------><------>init.name = list->name;
<------><------>init.ops = &lgm_clk_ddiv_ops;
<------><------>init.flags = list->flags;
<------><------>init.parent_data = list->parent_data;
<------><------>init.num_parents = 1;
<------><------>ddiv->membase = ctx->membase;
<------><------>ddiv->lock = ctx->lock;
<------><------>ddiv->reg = list->reg;
<------><------>ddiv->shift0 = list->shift0;
<------><------>ddiv->width0 = list->width0;
<------><------>ddiv->shift1 = list->shift1;
<------><------>ddiv->width1 = list->width1;
<------><------>ddiv->shift_gate = list->shift_gate;
<------><------>ddiv->width_gate = list->width_gate;
<------><------>ddiv->shift2 = list->ex_shift;
<------><------>ddiv->width2 = list->ex_width;
<------><------>ddiv->flags = list->div_flags;
<------><------>ddiv->mult = 2;
<------><------>ddiv->div = 5;
<------><------>ddiv->hw.init = &init;
<------><------>hw = &ddiv->hw;
<------><------>ret = devm_clk_hw_register(dev, hw);
<------><------>if (ret) {
<------><------><------>dev_err(dev, "register clk: %s failed!\n", list->name);
<------><------><------>return ret;
<------><------>}
<------><------>ctx->clk_data.hws[list->id] = hw;
<------>}
<------>return 0;
}