linux/drivers/media/v4l2-core/v4l2-clk.c
<<
>>
Prefs
   1/*
   2 * V4L2 clock service
   3 *
   4 * Copyright (C) 2012-2013, Guennadi Liakhovetski <g.liakhovetski@gmx.de>
   5 *
   6 * This program is free software; you can redistribute it and/or modify
   7 * it under the terms of the GNU General Public License version 2 as
   8 * published by the Free Software Foundation.
   9 */
  10
  11#include <linux/atomic.h>
  12#include <linux/clk.h>
  13#include <linux/device.h>
  14#include <linux/errno.h>
  15#include <linux/list.h>
  16#include <linux/module.h>
  17#include <linux/mutex.h>
  18#include <linux/of.h>
  19#include <linux/slab.h>
  20#include <linux/string.h>
  21
  22#include <media/v4l2-clk.h>
  23#include <media/v4l2-subdev.h>
  24
  25static DEFINE_MUTEX(clk_lock);
  26static LIST_HEAD(clk_list);
  27
  28static struct v4l2_clk *v4l2_clk_find(const char *dev_id)
  29{
  30        struct v4l2_clk *clk;
  31
  32        list_for_each_entry(clk, &clk_list, list)
  33                if (!strcmp(dev_id, clk->dev_id))
  34                        return clk;
  35
  36        return ERR_PTR(-ENODEV);
  37}
  38
  39struct v4l2_clk *v4l2_clk_get(struct device *dev, const char *id)
  40{
  41        struct v4l2_clk *clk;
  42        struct clk *ccf_clk = clk_get(dev, id);
  43        char clk_name[V4L2_CLK_NAME_SIZE];
  44
  45        if (PTR_ERR(ccf_clk) == -EPROBE_DEFER)
  46                return ERR_PTR(-EPROBE_DEFER);
  47
  48        if (!IS_ERR_OR_NULL(ccf_clk)) {
  49                clk = kzalloc(sizeof(*clk), GFP_KERNEL);
  50                if (!clk) {
  51                        clk_put(ccf_clk);
  52                        return ERR_PTR(-ENOMEM);
  53                }
  54                clk->clk = ccf_clk;
  55
  56                return clk;
  57        }
  58
  59        mutex_lock(&clk_lock);
  60        clk = v4l2_clk_find(dev_name(dev));
  61
  62        /* if dev_name is not found, try use the OF name to find again  */
  63        if (PTR_ERR(clk) == -ENODEV && dev->of_node) {
  64                v4l2_clk_name_of(clk_name, sizeof(clk_name), dev->of_node);
  65                clk = v4l2_clk_find(clk_name);
  66        }
  67
  68        if (!IS_ERR(clk))
  69                atomic_inc(&clk->use_count);
  70        mutex_unlock(&clk_lock);
  71
  72        return clk;
  73}
  74EXPORT_SYMBOL(v4l2_clk_get);
  75
  76void v4l2_clk_put(struct v4l2_clk *clk)
  77{
  78        struct v4l2_clk *tmp;
  79
  80        if (IS_ERR(clk))
  81                return;
  82
  83        if (clk->clk) {
  84                clk_put(clk->clk);
  85                kfree(clk);
  86                return;
  87        }
  88
  89        mutex_lock(&clk_lock);
  90
  91        list_for_each_entry(tmp, &clk_list, list)
  92                if (tmp == clk)
  93                        atomic_dec(&clk->use_count);
  94
  95        mutex_unlock(&clk_lock);
  96}
  97EXPORT_SYMBOL(v4l2_clk_put);
  98
  99static int v4l2_clk_lock_driver(struct v4l2_clk *clk)
 100{
 101        struct v4l2_clk *tmp;
 102        int ret = -ENODEV;
 103
 104        mutex_lock(&clk_lock);
 105
 106        list_for_each_entry(tmp, &clk_list, list)
 107                if (tmp == clk) {
 108                        ret = !try_module_get(clk->ops->owner);
 109                        if (ret)
 110                                ret = -EFAULT;
 111                        break;
 112                }
 113
 114        mutex_unlock(&clk_lock);
 115
 116        return ret;
 117}
 118
 119static void v4l2_clk_unlock_driver(struct v4l2_clk *clk)
 120{
 121        module_put(clk->ops->owner);
 122}
 123
 124int v4l2_clk_enable(struct v4l2_clk *clk)
 125{
 126        int ret;
 127
 128        if (clk->clk)
 129                return clk_prepare_enable(clk->clk);
 130
 131        ret = v4l2_clk_lock_driver(clk);
 132        if (ret < 0)
 133                return ret;
 134
 135        mutex_lock(&clk->lock);
 136
 137        if (++clk->enable == 1 && clk->ops->enable) {
 138                ret = clk->ops->enable(clk);
 139                if (ret < 0)
 140                        clk->enable--;
 141        }
 142
 143        mutex_unlock(&clk->lock);
 144
 145        return ret;
 146}
 147EXPORT_SYMBOL(v4l2_clk_enable);
 148
 149/*
 150 * You might Oops if you try to disabled a disabled clock, because then the
 151 * driver isn't locked and could have been unloaded by now, so, don't do that
 152 */
 153void v4l2_clk_disable(struct v4l2_clk *clk)
 154{
 155        int enable;
 156
 157        if (clk->clk)
 158                return clk_disable_unprepare(clk->clk);
 159
 160        mutex_lock(&clk->lock);
 161
 162        enable = --clk->enable;
 163        if (WARN(enable < 0, "Unbalanced %s() on %s!\n", __func__,
 164                 clk->dev_id))
 165                clk->enable++;
 166        else if (!enable && clk->ops->disable)
 167                clk->ops->disable(clk);
 168
 169        mutex_unlock(&clk->lock);
 170
 171        v4l2_clk_unlock_driver(clk);
 172}
 173EXPORT_SYMBOL(v4l2_clk_disable);
 174
 175unsigned long v4l2_clk_get_rate(struct v4l2_clk *clk)
 176{
 177        int ret;
 178
 179        if (clk->clk)
 180                return clk_get_rate(clk->clk);
 181
 182        ret = v4l2_clk_lock_driver(clk);
 183        if (ret < 0)
 184                return ret;
 185
 186        mutex_lock(&clk->lock);
 187        if (!clk->ops->get_rate)
 188                ret = -ENOSYS;
 189        else
 190                ret = clk->ops->get_rate(clk);
 191        mutex_unlock(&clk->lock);
 192
 193        v4l2_clk_unlock_driver(clk);
 194
 195        return ret;
 196}
 197EXPORT_SYMBOL(v4l2_clk_get_rate);
 198
 199int v4l2_clk_set_rate(struct v4l2_clk *clk, unsigned long rate)
 200{
 201        int ret;
 202
 203        if (clk->clk) {
 204                long r = clk_round_rate(clk->clk, rate);
 205                if (r < 0)
 206                        return r;
 207                return clk_set_rate(clk->clk, r);
 208        }
 209
 210        ret = v4l2_clk_lock_driver(clk);
 211
 212        if (ret < 0)
 213                return ret;
 214
 215        mutex_lock(&clk->lock);
 216        if (!clk->ops->set_rate)
 217                ret = -ENOSYS;
 218        else
 219                ret = clk->ops->set_rate(clk, rate);
 220        mutex_unlock(&clk->lock);
 221
 222        v4l2_clk_unlock_driver(clk);
 223
 224        return ret;
 225}
 226EXPORT_SYMBOL(v4l2_clk_set_rate);
 227
 228struct v4l2_clk *v4l2_clk_register(const struct v4l2_clk_ops *ops,
 229                                   const char *dev_id,
 230                                   void *priv)
 231{
 232        struct v4l2_clk *clk;
 233        int ret;
 234
 235        if (!ops || !dev_id)
 236                return ERR_PTR(-EINVAL);
 237
 238        clk = kzalloc(sizeof(struct v4l2_clk), GFP_KERNEL);
 239        if (!clk)
 240                return ERR_PTR(-ENOMEM);
 241
 242        clk->dev_id = kstrdup(dev_id, GFP_KERNEL);
 243        if (!clk->dev_id) {
 244                ret = -ENOMEM;
 245                goto ealloc;
 246        }
 247        clk->ops = ops;
 248        clk->priv = priv;
 249        atomic_set(&clk->use_count, 0);
 250        mutex_init(&clk->lock);
 251
 252        mutex_lock(&clk_lock);
 253        if (!IS_ERR(v4l2_clk_find(dev_id))) {
 254                mutex_unlock(&clk_lock);
 255                ret = -EEXIST;
 256                goto eexist;
 257        }
 258        list_add_tail(&clk->list, &clk_list);
 259        mutex_unlock(&clk_lock);
 260
 261        return clk;
 262
 263eexist:
 264ealloc:
 265        kfree(clk->dev_id);
 266        kfree(clk);
 267        return ERR_PTR(ret);
 268}
 269EXPORT_SYMBOL(v4l2_clk_register);
 270
 271void v4l2_clk_unregister(struct v4l2_clk *clk)
 272{
 273        if (WARN(atomic_read(&clk->use_count),
 274                 "%s(): Refusing to unregister ref-counted %s clock!\n",
 275                 __func__, clk->dev_id))
 276                return;
 277
 278        mutex_lock(&clk_lock);
 279        list_del(&clk->list);
 280        mutex_unlock(&clk_lock);
 281
 282        kfree(clk->dev_id);
 283        kfree(clk);
 284}
 285EXPORT_SYMBOL(v4l2_clk_unregister);
 286
 287struct v4l2_clk_fixed {
 288        unsigned long rate;
 289        struct v4l2_clk_ops ops;
 290};
 291
 292static unsigned long fixed_get_rate(struct v4l2_clk *clk)
 293{
 294        struct v4l2_clk_fixed *priv = clk->priv;
 295        return priv->rate;
 296}
 297
 298struct v4l2_clk *__v4l2_clk_register_fixed(const char *dev_id,
 299                                unsigned long rate, struct module *owner)
 300{
 301        struct v4l2_clk *clk;
 302        struct v4l2_clk_fixed *priv = kzalloc(sizeof(*priv), GFP_KERNEL);
 303
 304        if (!priv)
 305                return ERR_PTR(-ENOMEM);
 306
 307        priv->rate = rate;
 308        priv->ops.get_rate = fixed_get_rate;
 309        priv->ops.owner = owner;
 310
 311        clk = v4l2_clk_register(&priv->ops, dev_id, priv);
 312        if (IS_ERR(clk))
 313                kfree(priv);
 314
 315        return clk;
 316}
 317EXPORT_SYMBOL(__v4l2_clk_register_fixed);
 318
 319void v4l2_clk_unregister_fixed(struct v4l2_clk *clk)
 320{
 321        kfree(clk->priv);
 322        v4l2_clk_unregister(clk);
 323}
 324EXPORT_SYMBOL(v4l2_clk_unregister_fixed);
 325