| |
| |
| |
| |
| |
| |
| #include <linux/clk-provider.h> |
| #include <linux/device.h> |
| #include <linux/of.h> |
| |
| #include "clk-cgu.h" |
| |
| #define GATE_HW_REG_STAT(reg) ((reg) + 0x0) |
| #define GATE_HW_REG_EN(reg) ((reg) + 0x4) |
| #define GATE_HW_REG_DIS(reg) ((reg) + 0x8) |
| #define MAX_DDIV_REG 8 |
| #define MAX_DIVIDER_VAL 64 |
| |
| #define to_lgm_clk_mux(_hw) container_of(_hw, struct lgm_clk_mux, hw) |
| #define to_lgm_clk_divider(_hw) container_of(_hw, struct lgm_clk_divider, hw) |
| #define to_lgm_clk_gate(_hw) container_of(_hw, struct lgm_clk_gate, hw) |
| #define to_lgm_clk_ddiv(_hw) container_of(_hw, struct lgm_clk_ddiv, hw) |
| |
| static struct clk_hw *lgm_clk_register_fixed(struct lgm_clk_provider *ctx, |
| <------><------><------><------><------> const struct lgm_clk_branch *list) |
| { |
| <------>unsigned long flags; |
| |
| <------>if (list->div_flags & CLOCK_FLAG_VAL_INIT) { |
| <------><------>spin_lock_irqsave(&ctx->lock, flags); |
| <------><------>lgm_set_clk_val(ctx->membase, list->div_off, list->div_shift, |
| <------><------><------><------>list->div_width, list->div_val); |
| <------><------>spin_unlock_irqrestore(&ctx->lock, flags); |
| <------>} |
| |
| <------>return clk_hw_register_fixed_rate(NULL, list->name, |
| <------><------><------><------><------> list->parent_data[0].name, |
| <------><------><------><------><------> list->flags, list->mux_flags); |
| } |
| |
| static u8 lgm_clk_mux_get_parent(struct clk_hw *hw) |
| { |
| <------>struct lgm_clk_mux *mux = to_lgm_clk_mux(hw); |
| <------>unsigned long flags; |
| <------>u32 val; |
| |
| <------>spin_lock_irqsave(&mux->lock, flags); |
| <------>if (mux->flags & MUX_CLK_SW) |
| <------><------>val = mux->reg; |
| <------>else |
| <------><------>val = lgm_get_clk_val(mux->membase, mux->reg, mux->shift, |
| <------><------><------><------> mux->width); |
| <------>spin_unlock_irqrestore(&mux->lock, flags); |
| <------>return clk_mux_val_to_index(hw, NULL, mux->flags, val); |
| } |
| |
| static int lgm_clk_mux_set_parent(struct clk_hw *hw, u8 index) |
| { |
| <------>struct lgm_clk_mux *mux = to_lgm_clk_mux(hw); |
| <------>unsigned long flags; |
| <------>u32 val; |
| |
| <------>val = clk_mux_index_to_val(NULL, mux->flags, index); |
| <------>spin_lock_irqsave(&mux->lock, flags); |
| <------>if (mux->flags & MUX_CLK_SW) |
| <------><------>mux->reg = val; |
| <------>else |
| <------><------>lgm_set_clk_val(mux->membase, mux->reg, mux->shift, |
| <------><------><------><------>mux->width, val); |
| <------>spin_unlock_irqrestore(&mux->lock, flags); |
| |
| <------>return 0; |
| } |
| |
| static int lgm_clk_mux_determine_rate(struct clk_hw *hw, |
| <------><------><------><------> struct clk_rate_request *req) |
| { |
| <------>struct lgm_clk_mux *mux = to_lgm_clk_mux(hw); |
| |
| <------>return clk_mux_determine_rate_flags(hw, req, mux->flags); |
| } |
| |
| static const struct clk_ops lgm_clk_mux_ops = { |
| <------>.get_parent = lgm_clk_mux_get_parent, |
| <------>.set_parent = lgm_clk_mux_set_parent, |
| <------>.determine_rate = lgm_clk_mux_determine_rate, |
| }; |
| |
| static struct clk_hw * |
| lgm_clk_register_mux(struct lgm_clk_provider *ctx, |
| <------><------> const struct lgm_clk_branch *list) |
| { |
| <------>unsigned long flags, cflags = list->mux_flags; |
| <------>struct device *dev = ctx->dev; |
| <------>u8 shift = list->mux_shift; |
| <------>u8 width = list->mux_width; |
| <------>struct clk_init_data init = {}; |
| <------>struct lgm_clk_mux *mux; |
| <------>u32 reg = list->mux_off; |
| <------>struct clk_hw *hw; |
| <------>int ret; |
| |
| <------>mux = devm_kzalloc(dev, sizeof(*mux), GFP_KERNEL); |
| <------>if (!mux) |
| <------><------>return ERR_PTR(-ENOMEM); |
| |
| <------>init.name = list->name; |
| <------>init.ops = &lgm_clk_mux_ops; |
| <------>init.flags = list->flags; |
| <------>init.parent_data = list->parent_data; |
| <------>init.num_parents = list->num_parents; |
| |
| <------>mux->membase = ctx->membase; |
| <------>mux->lock = ctx->lock; |
| <------>mux->reg = reg; |
| <------>mux->shift = shift; |
| <------>mux->width = width; |
| <------>mux->flags = cflags; |
| <------>mux->hw.init = &init; |
| |
| <------>hw = &mux->hw; |
| <------>ret = devm_clk_hw_register(dev, hw); |
| <------>if (ret) |
| <------><------>return ERR_PTR(ret); |
| |
| <------>if (cflags & CLOCK_FLAG_VAL_INIT) { |
| <------><------>spin_lock_irqsave(&mux->lock, flags); |
| <------><------>lgm_set_clk_val(mux->membase, reg, shift, width, list->mux_val); |
| <------><------>spin_unlock_irqrestore(&mux->lock, flags); |
| <------>} |
| |
| <------>return hw; |
| } |
| |
| static unsigned long |
| lgm_clk_divider_recalc_rate(struct clk_hw *hw, unsigned long parent_rate) |
| { |
| <------>struct lgm_clk_divider *divider = to_lgm_clk_divider(hw); |
| <------>unsigned long flags; |
| <------>unsigned int val; |
| |
| <------>spin_lock_irqsave(÷r->lock, flags); |
| <------>val = lgm_get_clk_val(divider->membase, divider->reg, |
| <------><------><------> divider->shift, divider->width); |
| <------>spin_unlock_irqrestore(÷r->lock, flags); |
| |
| <------>return divider_recalc_rate(hw, parent_rate, val, divider->table, |
| <------><------><------><------> divider->flags, divider->width); |
| } |
| |
| static long |
| lgm_clk_divider_round_rate(struct clk_hw *hw, unsigned long rate, |
| <------><------><------> unsigned long *prate) |
| { |
| <------>struct lgm_clk_divider *divider = to_lgm_clk_divider(hw); |
| |
| <------>return divider_round_rate(hw, rate, prate, divider->table, |
| <------><------><------><------> divider->width, divider->flags); |
| } |
| |
| static int |
| lgm_clk_divider_set_rate(struct clk_hw *hw, unsigned long rate, |
| <------><------><------> unsigned long prate) |
| { |
| <------>struct lgm_clk_divider *divider = to_lgm_clk_divider(hw); |
| <------>unsigned long flags; |
| <------>int value; |
| |
| <------>value = divider_get_val(rate, prate, divider->table, |
| <------><------><------><------>divider->width, divider->flags); |
| <------>if (value < 0) |
| <------><------>return value; |
| |
| <------>spin_lock_irqsave(÷r->lock, flags); |
| <------>lgm_set_clk_val(divider->membase, divider->reg, |
| <------><------><------>divider->shift, divider->width, value); |
| <------>spin_unlock_irqrestore(÷r->lock, flags); |
| |
| <------>return 0; |
| } |
| |
| static int lgm_clk_divider_enable_disable(struct clk_hw *hw, int enable) |
| { |
| <------>struct lgm_clk_divider *div = to_lgm_clk_divider(hw); |
| <------>unsigned long flags; |
| |
| <------>spin_lock_irqsave(&div->lock, flags); |
| <------>lgm_set_clk_val(div->membase, div->reg, div->shift_gate, |
| <------><------><------>div->width_gate, enable); |
| <------>spin_unlock_irqrestore(&div->lock, flags); |
| <------>return 0; |
| } |
| |
| static int lgm_clk_divider_enable(struct clk_hw *hw) |
| { |
| <------>return lgm_clk_divider_enable_disable(hw, 1); |
| } |
| |
| static void lgm_clk_divider_disable(struct clk_hw *hw) |
| { |
| <------>lgm_clk_divider_enable_disable(hw, 0); |
| } |
| |
| static const struct clk_ops lgm_clk_divider_ops = { |
| <------>.recalc_rate = lgm_clk_divider_recalc_rate, |
| <------>.round_rate = lgm_clk_divider_round_rate, |
| <------>.set_rate = lgm_clk_divider_set_rate, |
| <------>.enable = lgm_clk_divider_enable, |
| <------>.disable = lgm_clk_divider_disable, |
| }; |
| |
| static struct clk_hw * |
| lgm_clk_register_divider(struct lgm_clk_provider *ctx, |
| <------><------><------> const struct lgm_clk_branch *list) |
| { |
| <------>unsigned long flags, cflags = list->div_flags; |
| <------>struct device *dev = ctx->dev; |
| <------>struct lgm_clk_divider *div; |
| <------>struct clk_init_data init = {}; |
| <------>u8 shift = list->div_shift; |
| <------>u8 width = list->div_width; |
| <------>u8 shift_gate = list->div_shift_gate; |
| <------>u8 width_gate = list->div_width_gate; |
| <------>u32 reg = list->div_off; |
| <------>struct clk_hw *hw; |
| <------>int ret; |
| |
| <------>div = devm_kzalloc(dev, sizeof(*div), GFP_KERNEL); |
| <------>if (!div) |
| <------><------>return ERR_PTR(-ENOMEM); |
| |
| <------>init.name = list->name; |
| <------>init.ops = &lgm_clk_divider_ops; |
| <------>init.flags = list->flags; |
| <------>init.parent_data = list->parent_data; |
| <------>init.num_parents = 1; |
| |
| <------>div->membase = ctx->membase; |
| <------>div->lock = ctx->lock; |
| <------>div->reg = reg; |
| <------>div->shift = shift; |
| <------>div->width = width; |
| <------>div->shift_gate = shift_gate; |
| <------>div->width_gate = width_gate; |
| <------>div->flags = cflags; |
| <------>div->table = list->div_table; |
| <------>div->hw.init = &init; |
| |
| <------>hw = &div->hw; |
| <------>ret = devm_clk_hw_register(dev, hw); |
| <------>if (ret) |
| <------><------>return ERR_PTR(ret); |
| |
| <------>if (cflags & CLOCK_FLAG_VAL_INIT) { |
| <------><------>spin_lock_irqsave(&div->lock, flags); |
| <------><------>lgm_set_clk_val(div->membase, reg, shift, width, list->div_val); |
| <------><------>spin_unlock_irqrestore(&div->lock, flags); |
| <------>} |
| |
| <------>return hw; |
| } |
| |
| static struct clk_hw * |
| lgm_clk_register_fixed_factor(struct lgm_clk_provider *ctx, |
| <------><------><------> const struct lgm_clk_branch *list) |
| { |
| <------>unsigned long flags; |
| <------>struct clk_hw *hw; |
| |
| <------>hw = clk_hw_register_fixed_factor(ctx->dev, list->name, |
| <------><------><------><------><------> list->parent_data[0].name, list->flags, |
| <------><------><------><------><------> list->mult, list->div); |
| <------>if (IS_ERR(hw)) |
| <------><------>return ERR_CAST(hw); |
| |
| <------>if (list->div_flags & CLOCK_FLAG_VAL_INIT) { |
| <------><------>spin_lock_irqsave(&ctx->lock, flags); |
| <------><------>lgm_set_clk_val(ctx->membase, list->div_off, list->div_shift, |
| <------><------><------><------>list->div_width, list->div_val); |
| <------><------>spin_unlock_irqrestore(&ctx->lock, flags); |
| <------>} |
| |
| <------>return hw; |
| } |
| |
| static int lgm_clk_gate_enable(struct clk_hw *hw) |
| { |
| <------>struct lgm_clk_gate *gate = to_lgm_clk_gate(hw); |
| <------>unsigned long flags; |
| <------>unsigned int reg; |
| |
| <------>spin_lock_irqsave(&gate->lock, flags); |
| <------>reg = GATE_HW_REG_EN(gate->reg); |
| <------>lgm_set_clk_val(gate->membase, reg, gate->shift, 1, 1); |
| <------>spin_unlock_irqrestore(&gate->lock, flags); |
| |
| <------>return 0; |
| } |
| |
| static void lgm_clk_gate_disable(struct clk_hw *hw) |
| { |
| <------>struct lgm_clk_gate *gate = to_lgm_clk_gate(hw); |
| <------>unsigned long flags; |
| <------>unsigned int reg; |
| |
| <------>spin_lock_irqsave(&gate->lock, flags); |
| <------>reg = GATE_HW_REG_DIS(gate->reg); |
| <------>lgm_set_clk_val(gate->membase, reg, gate->shift, 1, 1); |
| <------>spin_unlock_irqrestore(&gate->lock, flags); |
| } |
| |
| static int lgm_clk_gate_is_enabled(struct clk_hw *hw) |
| { |
| <------>struct lgm_clk_gate *gate = to_lgm_clk_gate(hw); |
| <------>unsigned int reg, ret; |
| <------>unsigned long flags; |
| |
| <------>spin_lock_irqsave(&gate->lock, flags); |
| <------>reg = GATE_HW_REG_STAT(gate->reg); |
| <------>ret = lgm_get_clk_val(gate->membase, reg, gate->shift, 1); |
| <------>spin_unlock_irqrestore(&gate->lock, flags); |
| |
| <------>return ret; |
| } |
| |
| static const struct clk_ops lgm_clk_gate_ops = { |
| <------>.enable = lgm_clk_gate_enable, |
| <------>.disable = lgm_clk_gate_disable, |
| <------>.is_enabled = lgm_clk_gate_is_enabled, |
| }; |
| |
| static struct clk_hw * |
| lgm_clk_register_gate(struct lgm_clk_provider *ctx, |
| <------><------> const struct lgm_clk_branch *list) |
| { |
| <------>unsigned long flags, cflags = list->gate_flags; |
| <------>const char *pname = list->parent_data[0].name; |
| <------>struct device *dev = ctx->dev; |
| <------>u8 shift = list->gate_shift; |
| <------>struct clk_init_data init = {}; |
| <------>struct lgm_clk_gate *gate; |
| <------>u32 reg = list->gate_off; |
| <------>struct clk_hw *hw; |
| <------>int ret; |
| |
| <------>gate = devm_kzalloc(dev, sizeof(*gate), GFP_KERNEL); |
| <------>if (!gate) |
| <------><------>return ERR_PTR(-ENOMEM); |
| |
| <------>init.name = list->name; |
| <------>init.ops = &lgm_clk_gate_ops; |
| <------>init.flags = list->flags; |
| <------>init.parent_names = pname ? &pname : NULL; |
| <------>init.num_parents = pname ? 1 : 0; |
| |
| <------>gate->membase = ctx->membase; |
| <------>gate->lock = ctx->lock; |
| <------>gate->reg = reg; |
| <------>gate->shift = shift; |
| <------>gate->flags = cflags; |
| <------>gate->hw.init = &init; |
| |
| <------>hw = &gate->hw; |
| <------>ret = devm_clk_hw_register(dev, hw); |
| <------>if (ret) |
| <------><------>return ERR_PTR(ret); |
| |
| <------>if (cflags & CLOCK_FLAG_VAL_INIT) { |
| <------><------>spin_lock_irqsave(&gate->lock, flags); |
| <------><------>lgm_set_clk_val(gate->membase, reg, shift, 1, list->gate_val); |
| <------><------>spin_unlock_irqrestore(&gate->lock, flags); |
| <------>} |
| |
| <------>return hw; |
| } |
| |
| int lgm_clk_register_branches(struct lgm_clk_provider *ctx, |
| <------><------><------> const struct lgm_clk_branch *list, |
| <------><------><------> unsigned int nr_clk) |
| { |
| <------>struct clk_hw *hw; |
| <------>unsigned int idx; |
| |
| <------>for (idx = 0; idx < nr_clk; idx++, list++) { |
| <------><------>switch (list->type) { |
| <------><------>case CLK_TYPE_FIXED: |
| <------><------><------>hw = lgm_clk_register_fixed(ctx, list); |
| <------><------><------>break; |
| <------><------>case CLK_TYPE_MUX: |
| <------><------><------>hw = lgm_clk_register_mux(ctx, list); |
| <------><------><------>break; |
| <------><------>case CLK_TYPE_DIVIDER: |
| <------><------><------>hw = lgm_clk_register_divider(ctx, list); |
| <------><------><------>break; |
| <------><------>case CLK_TYPE_FIXED_FACTOR: |
| <------><------><------>hw = lgm_clk_register_fixed_factor(ctx, list); |
| <------><------><------>break; |
| <------><------>case CLK_TYPE_GATE: |
| <------><------><------>hw = lgm_clk_register_gate(ctx, list); |
| <------><------><------>break; |
| <------><------>default: |
| <------><------><------>dev_err(ctx->dev, "invalid clk type\n"); |
| <------><------><------>return -EINVAL; |
| <------><------>} |
| |
| <------><------>if (IS_ERR(hw)) { |
| <------><------><------>dev_err(ctx->dev, |
| <------><------><------><------>"register clk: %s, type: %u failed!\n", |
| <------><------><------><------>list->name, list->type); |
| <------><------><------>return -EIO; |
| <------><------>} |
| <------><------>ctx->clk_data.hws[list->id] = hw; |
| <------>} |
| |
| <------>return 0; |
| } |
| |
| static unsigned long |
| lgm_clk_ddiv_recalc_rate(struct clk_hw *hw, unsigned long parent_rate) |
| { |
| <------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw); |
| <------>unsigned int div0, div1, exdiv; |
| <------>u64 prate; |
| |
| <------>div0 = lgm_get_clk_val(ddiv->membase, ddiv->reg, |
| <------><------><------> ddiv->shift0, ddiv->width0) + 1; |
| <------>div1 = lgm_get_clk_val(ddiv->membase, ddiv->reg, |
| <------><------><------> ddiv->shift1, ddiv->width1) + 1; |
| <------>exdiv = lgm_get_clk_val(ddiv->membase, ddiv->reg, |
| <------><------><------><------>ddiv->shift2, ddiv->width2); |
| <------>prate = (u64)parent_rate; |
| <------>do_div(prate, div0); |
| <------>do_div(prate, div1); |
| |
| <------>if (exdiv) { |
| <------><------>do_div(prate, ddiv->div); |
| <------><------>prate *= ddiv->mult; |
| <------>} |
| |
| <------>return prate; |
| } |
| |
| static int lgm_clk_ddiv_enable(struct clk_hw *hw) |
| { |
| <------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw); |
| <------>unsigned long flags; |
| |
| <------>spin_lock_irqsave(&ddiv->lock, flags); |
| <------>lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift_gate, |
| <------><------><------>ddiv->width_gate, 1); |
| <------>spin_unlock_irqrestore(&ddiv->lock, flags); |
| <------>return 0; |
| } |
| |
| static void lgm_clk_ddiv_disable(struct clk_hw *hw) |
| { |
| <------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw); |
| <------>unsigned long flags; |
| |
| <------>spin_lock_irqsave(&ddiv->lock, flags); |
| <------>lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift_gate, |
| <------><------><------>ddiv->width_gate, 0); |
| <------>spin_unlock_irqrestore(&ddiv->lock, flags); |
| } |
| |
| static int |
| lgm_clk_get_ddiv_val(u32 div, u32 *ddiv1, u32 *ddiv2) |
| { |
| <------>u32 idx, temp; |
| |
| <------>*ddiv1 = 1; |
| <------>*ddiv2 = 1; |
| |
| <------>if (div > MAX_DIVIDER_VAL) |
| <------><------>div = MAX_DIVIDER_VAL; |
| |
| <------>if (div > 1) { |
| <------><------>for (idx = 2; idx <= MAX_DDIV_REG; idx++) { |
| <------><------><------>temp = DIV_ROUND_UP_ULL((u64)div, idx); |
| <------><------><------>if (div % idx == 0 && temp <= MAX_DDIV_REG) |
| <------><------><------><------>break; |
| <------><------>} |
| |
| <------><------>if (idx > MAX_DDIV_REG) |
| <------><------><------>return -EINVAL; |
| |
| <------><------>*ddiv1 = temp; |
| <------><------>*ddiv2 = idx; |
| <------>} |
| |
| <------>return 0; |
| } |
| |
| static int |
| lgm_clk_ddiv_set_rate(struct clk_hw *hw, unsigned long rate, |
| <------><------> unsigned long prate) |
| { |
| <------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw); |
| <------>u32 div, ddiv1, ddiv2; |
| <------>unsigned long flags; |
| |
| <------>div = DIV_ROUND_CLOSEST_ULL((u64)prate, rate); |
| |
| <------>spin_lock_irqsave(&ddiv->lock, flags); |
| <------>if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) { |
| <------><------>div = DIV_ROUND_CLOSEST_ULL((u64)div, 5); |
| <------><------>div = div * 2; |
| <------>} |
| |
| <------>if (div <= 0) { |
| <------><------>spin_unlock_irqrestore(&ddiv->lock, flags); |
| <------><------>return -EINVAL; |
| <------>} |
| |
| <------>if (lgm_clk_get_ddiv_val(div, &ddiv1, &ddiv2)) { |
| <------><------>spin_unlock_irqrestore(&ddiv->lock, flags); |
| <------><------>return -EINVAL; |
| <------>} |
| |
| <------>lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift0, ddiv->width0, |
| <------><------><------>ddiv1 - 1); |
| |
| <------>lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift1, ddiv->width1, |
| <------><------><------>ddiv2 - 1); |
| <------>spin_unlock_irqrestore(&ddiv->lock, flags); |
| |
| <------>return 0; |
| } |
| |
| static long |
| lgm_clk_ddiv_round_rate(struct clk_hw *hw, unsigned long rate, |
| <------><------><------>unsigned long *prate) |
| { |
| <------>struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw); |
| <------>u32 div, ddiv1, ddiv2; |
| <------>unsigned long flags; |
| <------>u64 rate64; |
| |
| <------>div = DIV_ROUND_CLOSEST_ULL((u64)*prate, rate); |
| |
| <------> |
| <------>spin_lock_irqsave(&ddiv->lock, flags); |
| <------>if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) { |
| <------><------>div = div * 2; |
| <------><------>div = DIV_ROUND_CLOSEST_ULL((u64)div, 5); |
| <------>} |
| <------>spin_unlock_irqrestore(&ddiv->lock, flags); |
| |
| <------>if (div <= 0) |
| <------><------>return *prate; |
| |
| <------>if (lgm_clk_get_ddiv_val(div, &ddiv1, &ddiv2) != 0) |
| <------><------>if (lgm_clk_get_ddiv_val(div + 1, &ddiv1, &ddiv2) != 0) |
| <------><------><------>return -EINVAL; |
| |
| <------>rate64 = *prate; |
| <------>do_div(rate64, ddiv1); |
| <------>do_div(rate64, ddiv2); |
| |
| <------> |
| <------>spin_lock_irqsave(&ddiv->lock, flags); |
| <------>if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) { |
| <------><------>rate64 = rate64 * 2; |
| <------><------>rate64 = DIV_ROUND_CLOSEST_ULL(rate64, 5); |
| <------>} |
| <------>spin_unlock_irqrestore(&ddiv->lock, flags); |
| |
| <------>return rate64; |
| } |
| |
| static const struct clk_ops lgm_clk_ddiv_ops = { |
| <------>.recalc_rate = lgm_clk_ddiv_recalc_rate, |
| <------>.enable = lgm_clk_ddiv_enable, |
| <------>.disable = lgm_clk_ddiv_disable, |
| <------>.set_rate = lgm_clk_ddiv_set_rate, |
| <------>.round_rate = lgm_clk_ddiv_round_rate, |
| }; |
| |
| int lgm_clk_register_ddiv(struct lgm_clk_provider *ctx, |
| <------><------><------> const struct lgm_clk_ddiv_data *list, |
| <------><------><------> unsigned int nr_clk) |
| { |
| <------>struct device *dev = ctx->dev; |
| <------>struct clk_hw *hw; |
| <------>unsigned int idx; |
| <------>int ret; |
| |
| <------>for (idx = 0; idx < nr_clk; idx++, list++) { |
| <------><------>struct clk_init_data init = {}; |
| <------><------>struct lgm_clk_ddiv *ddiv; |
| |
| <------><------>ddiv = devm_kzalloc(dev, sizeof(*ddiv), GFP_KERNEL); |
| <------><------>if (!ddiv) |
| <------><------><------>return -ENOMEM; |
| |
| <------><------>init.name = list->name; |
| <------><------>init.ops = &lgm_clk_ddiv_ops; |
| <------><------>init.flags = list->flags; |
| <------><------>init.parent_data = list->parent_data; |
| <------><------>init.num_parents = 1; |
| |
| <------><------>ddiv->membase = ctx->membase; |
| <------><------>ddiv->lock = ctx->lock; |
| <------><------>ddiv->reg = list->reg; |
| <------><------>ddiv->shift0 = list->shift0; |
| <------><------>ddiv->width0 = list->width0; |
| <------><------>ddiv->shift1 = list->shift1; |
| <------><------>ddiv->width1 = list->width1; |
| <------><------>ddiv->shift_gate = list->shift_gate; |
| <------><------>ddiv->width_gate = list->width_gate; |
| <------><------>ddiv->shift2 = list->ex_shift; |
| <------><------>ddiv->width2 = list->ex_width; |
| <------><------>ddiv->flags = list->div_flags; |
| <------><------>ddiv->mult = 2; |
| <------><------>ddiv->div = 5; |
| <------><------>ddiv->hw.init = &init; |
| |
| <------><------>hw = &ddiv->hw; |
| <------><------>ret = devm_clk_hw_register(dev, hw); |
| <------><------>if (ret) { |
| <------><------><------>dev_err(dev, "register clk: %s failed!\n", list->name); |
| <------><------><------>return ret; |
| <------><------>} |
| <------><------>ctx->clk_data.hws[list->id] = hw; |
| <------>} |
| |
| <------>return 0; |
| } |
| |