// SPDX-License-Identifier: GPL-2.0 /* * Copyright 2017 NXP * * Dong Aisheng <aisheng.dong@nxp.com> */ #include <linux/clk.h> #include <linux/clk-provider.h> #include <linux/device.h> #include <linux/export.h> #include <linux/of.h> #include <linux/slab.h> static int __must_check of_clk_bulk_get(struct device_node *np, int num_clks, struct clk_bulk_data *clks) { int ret; int i; for (i = 0; i < num_clks; i++) { clks[i].id = NULL; clks[i].clk = NULL; } for (i = 0; i < num_clks; i++) { of_property_read_string_index(np, "clock-names", i, &clks[i].id); clks[i].clk = of_clk_get(np, i); if (IS_ERR(clks[i].clk)) { ret = PTR_ERR(clks[i].clk); pr_err("%pOF: Failed to get clk index: %d ret: %d\n", np, i, ret); clks[i].clk = NULL; goto err; } } return 0; err: clk_bulk_put(i, clks); return ret; } static int __must_check of_clk_bulk_get_all(struct device_node *np, struct clk_bulk_data **clks) { struct clk_bulk_data *clk_bulk; int num_clks; int ret; num_clks = of_clk_get_parent_count(np); if (!num_clks) return 0; clk_bulk = kmalloc_array(num_clks, sizeof(*clk_bulk), GFP_KERNEL); if (!clk_bulk) return -ENOMEM; ret = of_clk_bulk_get(np, num_clks, clk_bulk); if (ret) { kfree(clk_bulk); return ret; } *clks = clk_bulk; return num_clks; } void clk_bulk_put(int num_clks, struct clk_bulk_data *clks) { while (--num_clks >= 0) { clk_put(clks[num_clks].clk); clks[num_clks].clk = NULL; } } EXPORT_SYMBOL_GPL(clk_bulk_put); static int __clk_bulk_get(struct device *dev, int num_clks, struct clk_bulk_data *clks, bool optional) { int ret; int i; for (i = 0; i < num_clks; i++) clks[i].clk = NULL; for (i = 0; i < num_clks; i++) { clks[i].clk = clk_get(dev, clks[i].id); if (IS_ERR(clks[i].clk)) { ret = PTR_ERR(clks[i].clk); clks[i].clk = NULL; if (ret == -ENOENT && optional) continue; if (ret != -EPROBE_DEFER) dev_err(dev, "Failed to get clk '%s': %d\n", clks[i].id, ret); goto err; } } return 0; err: clk_bulk_put(i, clks); return ret; } int __must_check clk_bulk_get(struct device *dev, int num_clks, struct clk_bulk_data *clks) { return __clk_bulk_get(dev, num_clks, clks, false); } EXPORT_SYMBOL(clk_bulk_get); int __must_check clk_bulk_get_optional(struct device *dev, int num_clks, struct clk_bulk_data *clks) { return __clk_bulk_get(dev, num_clks, clks, true); } EXPORT_SYMBOL_GPL(clk_bulk_get_optional); void clk_bulk_put_all(int num_clks, struct clk_bulk_data *clks) { if (IS_ERR_OR_NULL(clks)) return; clk_bulk_put(num_clks, clks); kfree(clks); } EXPORT_SYMBOL(clk_bulk_put_all); int __must_check clk_bulk_get_all(struct device *dev, struct clk_bulk_data **clks) { struct device_node *np = dev_of_node(dev); if (!np) return 0; return of_clk_bulk_get_all(np, clks); } EXPORT_SYMBOL(clk_bulk_get_all); #ifdef CONFIG_HAVE_CLK_PREPARE /** * clk_bulk_unprepare - undo preparation of a set of clock sources * @num_clks: the number of clk_bulk_data * @clks: the clk_bulk_data table being unprepared * * clk_bulk_unprepare may sleep, which differentiates it from clk_bulk_disable. * Returns 0 on success, -EERROR otherwise. */ void clk_bulk_unprepare(int num_clks, const struct clk_bulk_data *clks) { while (--num_clks >= 0) clk_unprepare(clks[num_clks].clk); } EXPORT_SYMBOL_GPL(clk_bulk_unprepare); /** * clk_bulk_prepare - prepare a set of clocks * @num_clks: the number of clk_bulk_data * @clks: the clk_bulk_data table being prepared * * clk_bulk_prepare may sleep, which differentiates it from clk_bulk_enable. * Returns 0 on success, -EERROR otherwise. */ int __must_check clk_bulk_prepare(int num_clks, const struct clk_bulk_data *clks) { int ret; int i; for (i = 0; i < num_clks; i++) { ret = clk_prepare(clks[i].clk); if (ret) { pr_err("Failed to prepare clk '%s': %d\n", clks[i].id, ret); goto err; } } return 0; err: clk_bulk_unprepare(i, clks); return ret; } EXPORT_SYMBOL_GPL(clk_bulk_prepare); #endif /* CONFIG_HAVE_CLK_PREPARE */ /** * clk_bulk_disable - gate a set of clocks * @num_clks: the number of clk_bulk_data * @clks: the clk_bulk_data table being gated * * clk_bulk_disable must not sleep, which differentiates it from * clk_bulk_unprepare. clk_bulk_disable must be called before * clk_bulk_unprepare. */ void clk_bulk_disable(int num_clks, const struct clk_bulk_data *clks) { while (--num_clks >= 0) clk_disable(clks[num_clks].clk); } EXPORT_SYMBOL_GPL(clk_bulk_disable); /** * clk_bulk_enable - ungate a set of clocks * @num_clks: the number of clk_bulk_data * @clks: the clk_bulk_data table being ungated * * clk_bulk_enable must not sleep * Returns 0 on success, -EERROR otherwise. */ int __must_check clk_bulk_enable(int num_clks, const struct clk_bulk_data *clks) { int ret; int i; for (i = 0; i < num_clks; i++) { ret = clk_enable(clks[i].clk); if (ret) { pr_err("Failed to enable clk '%s': %d\n", clks[i].id, ret); goto err; } } return 0; err: clk_bulk_disable(i, clks); return ret; } EXPORT_SYMBOL_GPL(clk_bulk_enable);