linux/drivers/clk/mvebu/clk-cpu.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Marvell MVEBU CPU clock handling.
   4 *
   5 * Copyright (C) 2012 Marvell
   6 *
   7 * Gregory CLEMENT <gregory.clement@free-electrons.com>
   8 *
   9 */
  10#include <linux/kernel.h>
  11#include <linux/slab.h>
  12#include <linux/clk.h>
  13#include <linux/clk-provider.h>
  14#include <linux/of_address.h>
  15#include <linux/io.h>
  16#include <linux/of.h>
  17#include <linux/delay.h>
  18#include <linux/mvebu-pmsu.h>
  19#include <asm/smp_plat.h>
  20
  21#define SYS_CTRL_CLK_DIVIDER_CTRL_OFFSET               0x0
  22#define   SYS_CTRL_CLK_DIVIDER_CTRL_RESET_ALL          0xff
  23#define   SYS_CTRL_CLK_DIVIDER_CTRL_RESET_SHIFT        8
  24#define SYS_CTRL_CLK_DIVIDER_CTRL2_OFFSET              0x8
  25#define   SYS_CTRL_CLK_DIVIDER_CTRL2_NBCLK_RATIO_SHIFT 16
  26#define SYS_CTRL_CLK_DIVIDER_VALUE_OFFSET              0xC
  27#define SYS_CTRL_CLK_DIVIDER_MASK                      0x3F
  28
  29#define PMU_DFS_RATIO_SHIFT 16
  30#define PMU_DFS_RATIO_MASK  0x3F
  31
  32#define MAX_CPU     4
  33struct cpu_clk {
  34        struct clk_hw hw;
  35        int cpu;
  36        const char *clk_name;
  37        const char *parent_name;
  38        void __iomem *reg_base;
  39        void __iomem *pmu_dfs;
  40};
  41
  42static struct clk **clks;
  43
  44static struct clk_onecell_data clk_data;
  45
  46#define to_cpu_clk(p) container_of(p, struct cpu_clk, hw)
  47
  48static unsigned long clk_cpu_recalc_rate(struct clk_hw *hwclk,
  49                                         unsigned long parent_rate)
  50{
  51        struct cpu_clk *cpuclk = to_cpu_clk(hwclk);
  52        u32 reg, div;
  53
  54        reg = readl(cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_VALUE_OFFSET);
  55        div = (reg >> (cpuclk->cpu * 8)) & SYS_CTRL_CLK_DIVIDER_MASK;
  56        return parent_rate / div;
  57}
  58
  59static long clk_cpu_round_rate(struct clk_hw *hwclk, unsigned long rate,
  60                               unsigned long *parent_rate)
  61{
  62        /* Valid ratio are 1:1, 1:2 and 1:3 */
  63        u32 div;
  64
  65        div = *parent_rate / rate;
  66        if (div == 0)
  67                div = 1;
  68        else if (div > 3)
  69                div = 3;
  70
  71        return *parent_rate / div;
  72}
  73
  74static int clk_cpu_off_set_rate(struct clk_hw *hwclk, unsigned long rate,
  75                                unsigned long parent_rate)
  76
  77{
  78        struct cpu_clk *cpuclk = to_cpu_clk(hwclk);
  79        u32 reg, div;
  80        u32 reload_mask;
  81
  82        div = parent_rate / rate;
  83        reg = (readl(cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_VALUE_OFFSET)
  84                & (~(SYS_CTRL_CLK_DIVIDER_MASK << (cpuclk->cpu * 8))))
  85                | (div << (cpuclk->cpu * 8));
  86        writel(reg, cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_VALUE_OFFSET);
  87        /* Set clock divider reload smooth bit mask */
  88        reload_mask = 1 << (20 + cpuclk->cpu);
  89
  90        reg = readl(cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_CTRL_OFFSET)
  91            | reload_mask;
  92        writel(reg, cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_CTRL_OFFSET);
  93
  94        /* Now trigger the clock update */
  95        reg = readl(cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_CTRL_OFFSET)
  96            | 1 << 24;
  97        writel(reg, cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_CTRL_OFFSET);
  98
  99        /* Wait for clocks to settle down then clear reload request */
 100        udelay(1000);
 101        reg &= ~(reload_mask | 1 << 24);
 102        writel(reg, cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_CTRL_OFFSET);
 103        udelay(1000);
 104
 105        return 0;
 106}
 107
 108static int clk_cpu_on_set_rate(struct clk_hw *hwclk, unsigned long rate,
 109                               unsigned long parent_rate)
 110{
 111        u32 reg;
 112        unsigned long fabric_div, target_div, cur_rate;
 113        struct cpu_clk *cpuclk = to_cpu_clk(hwclk);
 114
 115        /*
 116         * PMU DFS registers are not mapped, Device Tree does not
 117         * describes them. We cannot change the frequency dynamically.
 118         */
 119        if (!cpuclk->pmu_dfs)
 120                return -ENODEV;
 121
 122        cur_rate = clk_hw_get_rate(hwclk);
 123
 124        reg = readl(cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_CTRL2_OFFSET);
 125        fabric_div = (reg >> SYS_CTRL_CLK_DIVIDER_CTRL2_NBCLK_RATIO_SHIFT) &
 126                SYS_CTRL_CLK_DIVIDER_MASK;
 127
 128        /* Frequency is going up */
 129        if (rate == 2 * cur_rate)
 130                target_div = fabric_div / 2;
 131        /* Frequency is going down */
 132        else
 133                target_div = fabric_div;
 134
 135        if (target_div == 0)
 136                target_div = 1;
 137
 138        reg = readl(cpuclk->pmu_dfs);
 139        reg &= ~(PMU_DFS_RATIO_MASK << PMU_DFS_RATIO_SHIFT);
 140        reg |= (target_div << PMU_DFS_RATIO_SHIFT);
 141        writel(reg, cpuclk->pmu_dfs);
 142
 143        reg = readl(cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_CTRL_OFFSET);
 144        reg |= (SYS_CTRL_CLK_DIVIDER_CTRL_RESET_ALL <<
 145                SYS_CTRL_CLK_DIVIDER_CTRL_RESET_SHIFT);
 146        writel(reg, cpuclk->reg_base + SYS_CTRL_CLK_DIVIDER_CTRL_OFFSET);
 147
 148        return mvebu_pmsu_dfs_request(cpuclk->cpu);
 149}
 150
 151static int clk_cpu_set_rate(struct clk_hw *hwclk, unsigned long rate,
 152                            unsigned long parent_rate)
 153{
 154        if (__clk_is_enabled(hwclk->clk))
 155                return clk_cpu_on_set_rate(hwclk, rate, parent_rate);
 156        else
 157                return clk_cpu_off_set_rate(hwclk, rate, parent_rate);
 158}
 159
 160static const struct clk_ops cpu_ops = {
 161        .recalc_rate = clk_cpu_recalc_rate,
 162        .round_rate = clk_cpu_round_rate,
 163        .set_rate = clk_cpu_set_rate,
 164};
 165
 166static void __init of_cpu_clk_setup(struct device_node *node)
 167{
 168        struct cpu_clk *cpuclk;
 169        void __iomem *clock_complex_base = of_iomap(node, 0);
 170        void __iomem *pmu_dfs_base = of_iomap(node, 1);
 171        int ncpus = 0;
 172        struct device_node *dn;
 173
 174        if (clock_complex_base == NULL) {
 175                pr_err("%s: clock-complex base register not set\n",
 176                        __func__);
 177                return;
 178        }
 179
 180        if (pmu_dfs_base == NULL)
 181                pr_warn("%s: pmu-dfs base register not set, dynamic frequency scaling not available\n",
 182                        __func__);
 183
 184        for_each_of_cpu_node(dn)
 185                ncpus++;
 186
 187        cpuclk = kcalloc(ncpus, sizeof(*cpuclk), GFP_KERNEL);
 188        if (WARN_ON(!cpuclk))
 189                goto cpuclk_out;
 190
 191        clks = kcalloc(ncpus, sizeof(*clks), GFP_KERNEL);
 192        if (WARN_ON(!clks))
 193                goto clks_out;
 194
 195        for_each_of_cpu_node(dn) {
 196                struct clk_init_data init;
 197                struct clk *clk;
 198                char *clk_name = kzalloc(5, GFP_KERNEL);
 199                int cpu, err;
 200
 201                if (WARN_ON(!clk_name))
 202                        goto bail_out;
 203
 204                err = of_property_read_u32(dn, "reg", &cpu);
 205                if (WARN_ON(err))
 206                        goto bail_out;
 207
 208                sprintf(clk_name, "cpu%d", cpu);
 209
 210                cpuclk[cpu].parent_name = of_clk_get_parent_name(node, 0);
 211                cpuclk[cpu].clk_name = clk_name;
 212                cpuclk[cpu].cpu = cpu;
 213                cpuclk[cpu].reg_base = clock_complex_base;
 214                if (pmu_dfs_base)
 215                        cpuclk[cpu].pmu_dfs = pmu_dfs_base + 4 * cpu;
 216                cpuclk[cpu].hw.init = &init;
 217
 218                init.name = cpuclk[cpu].clk_name;
 219                init.ops = &cpu_ops;
 220                init.flags = 0;
 221                init.parent_names = &cpuclk[cpu].parent_name;
 222                init.num_parents = 1;
 223
 224                clk = clk_register(NULL, &cpuclk[cpu].hw);
 225                if (WARN_ON(IS_ERR(clk)))
 226                        goto bail_out;
 227                clks[cpu] = clk;
 228        }
 229        clk_data.clk_num = MAX_CPU;
 230        clk_data.clks = clks;
 231        of_clk_add_provider(node, of_clk_src_onecell_get, &clk_data);
 232
 233        return;
 234bail_out:
 235        kfree(clks);
 236        while(ncpus--)
 237                kfree(cpuclk[ncpus].clk_name);
 238clks_out:
 239        kfree(cpuclk);
 240cpuclk_out:
 241        iounmap(clock_complex_base);
 242}
 243
 244CLK_OF_DECLARE(armada_xp_cpu_clock, "marvell,armada-xp-cpu-clock",
 245                                         of_cpu_clk_setup);
 246
 247static void __init of_mv98dx3236_cpu_clk_setup(struct device_node *node)
 248{
 249        of_clk_add_provider(node, of_clk_src_simple_get, NULL);
 250}
 251
 252CLK_OF_DECLARE(mv98dx3236_cpu_clock, "marvell,mv98dx3236-cpu-clock",
 253                                         of_mv98dx3236_cpu_clk_setup);
 254