/* * Register map access API - MMIO support * * Copyright (c) 2012, NVIDIA CORPORATION. All rights reserved. * * This program is free software; you can redistribute it and/or modify it * under the terms and conditions of the GNU General Public License, * version 2, as published by the Free Software Foundation. * * This program is distributed in the hope it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for * more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see . */ #include #include #include #include #include #include struct regmap_mmio_context { void __iomem *regs; unsigned val_bytes; struct clk *clk; }; static int regmap_mmio_gather_write(void *context, const void *reg, size_t reg_size, const void *val, size_t val_size) { struct regmap_mmio_context *ctx = context; u32 offset; int ret; BUG_ON(reg_size != 4); if (!IS_ERR(ctx->clk)) { ret = clk_enable(ctx->clk); if (ret < 0) return ret; } offset = *(u32 *)reg; while (val_size) { switch (ctx->val_bytes) { case 1: writeb(*(u8 *)val, ctx->regs + offset); break; case 2: writew(*(u16 *)val, ctx->regs + offset); break; case 4: writel(*(u32 *)val, ctx->regs + offset); break; #ifdef CONFIG_64BIT case 8: writeq(*(u64 *)val, ctx->regs + offset); break; #endif default: /* Should be caught by regmap_mmio_check_config */ BUG(); } val_size -= ctx->val_bytes; val += ctx->val_bytes; offset += ctx->val_bytes; } if (!IS_ERR(ctx->clk)) clk_disable(ctx->clk); return 0; } static int regmap_mmio_write(void *context, const void *data, size_t count) { BUG_ON(count < 4); return regmap_mmio_gather_write(context, data, 4, data + 4, count - 4); } static int regmap_mmio_read(void *context, const void *reg, size_t reg_size, void *val, size_t val_size) { struct regmap_mmio_context *ctx = context; u32 offset; int ret; BUG_ON(reg_size != 4); if (!IS_ERR(ctx->clk)) { ret = clk_enable(ctx->clk); if (ret < 0) return ret; } offset = *(u32 *)reg; while (val_size) { switch (ctx->val_bytes) { case 1: *(u8 *)val = readb(ctx->regs + offset); break; case 2: *(u16 *)val = readw(ctx->regs + offset); break; case 4: *(u32 *)val = readl(ctx->regs + offset); break; #ifdef CONFIG_64BIT case 8: *(u64 *)val = readq(ctx->regs + offset); break; #endif default: /* Should be caught by regmap_mmio_check_config */ BUG(); } val_size -= ctx->val_bytes; val += ctx->val_bytes; offset += ctx->val_bytes; } if (!IS_ERR(ctx->clk)) clk_disable(ctx->clk); return 0; } static void regmap_mmio_free_context(void *context) { struct regmap_mmio_context *ctx = context; if (!IS_ERR(ctx->clk)) { clk_unprepare(ctx->clk); clk_put(ctx->clk); } kfree(context); } static struct regmap_bus regmap_mmio = { .fast_io = true, .write = regmap_mmio_write, .gather_write = regmap_mmio_gather_write, .read = regmap_mmio_read, .free_context = regmap_mmio_free_context, .reg_format_endian_default = REGMAP_ENDIAN_NATIVE, .val_format_endian_default = REGMAP_ENDIAN_NATIVE, }; static struct regmap_mmio_context *regmap_mmio_gen_context(struct device *dev, const char *clk_id, void __iomem *regs, const struct regmap_config *config) { struct regmap_mmio_context *ctx; int min_stride; int ret; if (config->reg_bits != 32) return ERR_PTR(-EINVAL); if (config->pad_bits) return ERR_PTR(-EINVAL); switch (config->val_bits) { case 8: /* The core treats 0 as 1 */ min_stride = 0; break; case 16: min_stride = 2; break; case 32: min_stride = 4; break; #ifdef CONFIG_64BIT case 64: min_stride = 8; break; #endif break; default: return ERR_PTR(-EINVAL); } if (config->reg_stride < min_stride) return ERR_PTR(-EINVAL); switch (config->reg_format_endian) { case REGMAP_ENDIAN_DEFAULT: case REGMAP_ENDIAN_NATIVE: break; default: return ERR_PTR(-EINVAL); } ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); if (!ctx) return ERR_PTR(-ENOMEM); ctx->regs = regs; ctx->val_bytes = config->val_bits / 8; ctx->clk = ERR_PTR(-ENODEV); ctx->clk = clk_get(dev, clk_id); if (!IS_ERR(ctx->clk)) { ret = clk_prepare(ctx->clk); if (ret < 0) { clk_put(ctx->clk); goto err_free; } } else { ctx->clk = NULL; } return ctx; err_free: kfree(ctx); return ERR_PTR(ret); } /** * regmap_init_mmio_clk(): Initialise register map with register clock * * @dev: Device that will be interacted with * @clk_id: register clock consumer ID * @regs: Pointer to memory-mapped IO region * @config: Configuration for register map * * The return value will be an ERR_PTR() on error or a valid pointer to * a struct regmap. */ struct regmap *regmap_init_mmio_clk(struct device *dev, const char *clk_id, void __iomem *regs, const struct regmap_config *config) { struct regmap_mmio_context *ctx; ctx = regmap_mmio_gen_context(dev, clk_id, regs, config); if (IS_ERR(ctx)) return ERR_CAST(ctx); return regmap_init(dev, ®map_mmio, ctx, config); } EXPORT_SYMBOL_GPL(regmap_init_mmio_clk); /** * devm_regmap_init_mmio_clk(): Initialise managed register map with clock * * @dev: Device that will be interacted with * @clk_id: register clock consumer ID * @regs: Pointer to memory-mapped IO region * @config: Configuration for register map * * The return value will be an ERR_PTR() on error or a valid pointer * to a struct regmap. The regmap will be automatically freed by the * device management code. */ struct regmap *devm_regmap_init_mmio_clk(struct device *dev, const char *clk_id, void __iomem *regs, const struct regmap_config *config) { struct regmap_mmio_context *ctx; ctx = regmap_mmio_gen_context(dev, clk_id, regs, config); if (IS_ERR(ctx)) return ERR_CAST(ctx); return devm_regmap_init(dev, ®map_mmio, ctx, config); } EXPORT_SYMBOL_GPL(devm_regmap_init_mmio_clk); MODULE_LICENSE("GPL v2");