#include <linux/pm_domain.h>
#include <linux/pm_runtime.h>
+enum sh_tmu_model {
+ SH_TMU_LEGACY,
+ SH_TMU,
+ SH_TMU_SH3,
+};
+
struct sh_tmu_device;
struct sh_tmu_channel {
void __iomem *mapbase;
struct clk *clk;
+ enum sh_tmu_model model;
+
struct sh_tmu_channel *channels;
unsigned int num_channels;
+
+ bool has_clockevent;
+ bool has_clocksource;
};
static DEFINE_RAW_SPINLOCK(sh_tmu_lock);
{
unsigned long offs;
- if (reg_nr == TSTR)
- return ioread8(ch->tmu->mapbase);
+ if (reg_nr == TSTR) {
+ switch (ch->tmu->model) {
+ case SH_TMU_LEGACY:
+ return ioread8(ch->tmu->mapbase);
+ case SH_TMU_SH3:
+ return ioread8(ch->tmu->mapbase + 2);
+ case SH_TMU:
+ return ioread8(ch->tmu->mapbase + 4);
+ }
+ }
offs = reg_nr << 2;
unsigned long offs;
if (reg_nr == TSTR) {
- iowrite8(value, ch->tmu->mapbase);
- return;
+ switch (ch->tmu->model) {
+ case SH_TMU_LEGACY:
+ return iowrite8(value, ch->tmu->mapbase);
+ case SH_TMU_SH3:
+ return iowrite8(value, ch->tmu->mapbase + 2);
+ case SH_TMU:
+ return iowrite8(value, ch->tmu->mapbase + 4);
+ }
}
offs = reg_nr << 2;
static int sh_tmu_register(struct sh_tmu_channel *ch, const char *name,
bool clockevent, bool clocksource)
{
- if (clockevent)
+ if (clockevent) {
+ ch->tmu->has_clockevent = true;
sh_tmu_register_clockevent(ch, name);
- else if (clocksource)
+ } else if (clocksource) {
+ ch->tmu->has_clocksource = true;
sh_tmu_register_clocksource(ch, name);
+ }
return 0;
}
-static int sh_tmu_channel_setup(struct sh_tmu_channel *ch,
+static int sh_tmu_channel_setup(struct sh_tmu_channel *ch, unsigned int index,
+ bool clockevent, bool clocksource,
struct sh_tmu_device *tmu)
{
- struct sh_timer_config *cfg = tmu->pdev->dev.platform_data;
+ /* Skip unused channels. */
+ if (!clockevent && !clocksource)
+ return 0;
ch->tmu = tmu;
- /*
- * The SH3 variant (SH770x, SH7705, SH7710 and SH7720) maps channel
- * registers blocks at base + 2 + 12 * index, while all other variants
- * map them at base + 4 + 12 * index. We can compute the index by just
- * dividing by 12, the 2 bytes or 4 bytes offset being hidden by the
- * integer division.
- */
- ch->index = cfg->channel_offset / 12;
+ if (tmu->model == SH_TMU_LEGACY) {
+ struct sh_timer_config *cfg = tmu->pdev->dev.platform_data;
+
+ /*
+ * The SH3 variant (SH770x, SH7705, SH7710 and SH7720) maps
+ * channel registers blocks at base + 2 + 12 * index, while all
+ * other variants map them at base + 4 + 12 * index. We can
+ * compute the index by just dividing by 12, the 2 bytes or 4
+ * bytes offset being hidden by the integer division.
+ */
+ ch->index = cfg->channel_offset / 12;
+ ch->base = tmu->mapbase + cfg->channel_offset;
+ } else {
+ ch->index = index;
+
+ if (tmu->model == SH_TMU_SH3)
+ ch->base = tmu->mapbase + 4 + ch->index * 12;
+ else
+ ch->base = tmu->mapbase + 8 + ch->index * 12;
+ }
- ch->irq = platform_get_irq(tmu->pdev, 0);
+ ch->irq = platform_get_irq(tmu->pdev, ch->index);
if (ch->irq < 0) {
dev_err(&tmu->pdev->dev, "ch%u: failed to get irq\n",
ch->index);
ch->enable_count = 0;
return sh_tmu_register(ch, dev_name(&tmu->pdev->dev),
- cfg->clockevent_rating != 0,
- cfg->clocksource_rating != 0);
+ clockevent, clocksource);
}
-static int sh_tmu_setup(struct sh_tmu_device *tmu, struct platform_device *pdev)
+static int sh_tmu_map_memory(struct sh_tmu_device *tmu)
{
- struct sh_timer_config *cfg = pdev->dev.platform_data;
struct resource *res;
- void __iomem *base;
- int ret;
- ret = -ENXIO;
-
- tmu->pdev = pdev;
-
- if (!cfg) {
- dev_err(&tmu->pdev->dev, "missing platform data\n");
- goto err0;
- }
-
- platform_set_drvdata(pdev, tmu);
res = platform_get_resource(tmu->pdev, IORESOURCE_MEM, 0);
if (!res) {
dev_err(&tmu->pdev->dev, "failed to get I/O memory\n");
- goto err0;
+ return -ENXIO;
}
+ tmu->mapbase = ioremap_nocache(res->start, resource_size(res));
+ if (tmu->mapbase == NULL)
+ return -ENXIO;
+
/*
- * Map memory, let base point to our channel and mapbase to the
- * start/stop shared register.
+ * In legacy platform device configuration (with one device per channel)
+ * the resource points to the channel base address.
*/
- base = ioremap_nocache(res->start, resource_size(res));
- if (base == NULL) {
- dev_err(&tmu->pdev->dev, "failed to remap I/O memory\n");
- goto err0;
+ if (tmu->model == SH_TMU_LEGACY) {
+ struct sh_timer_config *cfg = tmu->pdev->dev.platform_data;
+ tmu->mapbase -= cfg->channel_offset;
}
- tmu->mapbase = base - cfg->channel_offset;
+ return 0;
+}
- /* get hold of clock */
+static void sh_tmu_unmap_memory(struct sh_tmu_device *tmu)
+{
+ if (tmu->model == SH_TMU_LEGACY) {
+ struct sh_timer_config *cfg = tmu->pdev->dev.platform_data;
+ tmu->mapbase += cfg->channel_offset;
+ }
+
+ iounmap(tmu->mapbase);
+}
+
+static int sh_tmu_setup(struct sh_tmu_device *tmu, struct platform_device *pdev)
+{
+ struct sh_timer_config *cfg = pdev->dev.platform_data;
+ const struct platform_device_id *id = pdev->id_entry;
+ unsigned int i;
+ int ret;
+
+ if (!cfg) {
+ dev_err(&tmu->pdev->dev, "missing platform data\n");
+ return -ENXIO;
+ }
+
+ tmu->pdev = pdev;
+ tmu->model = id->driver_data;
+
+ /* Get hold of clock. */
tmu->clk = clk_get(&tmu->pdev->dev, "tmu_fck");
if (IS_ERR(tmu->clk)) {
dev_err(&tmu->pdev->dev, "cannot get clock\n");
- ret = PTR_ERR(tmu->clk);
- goto err1;
+ return PTR_ERR(tmu->clk);
}
ret = clk_prepare(tmu->clk);
if (ret < 0)
- goto err2;
+ goto err_clk_put;
+
+ /* Map the memory resource. */
+ ret = sh_tmu_map_memory(tmu);
+ if (ret < 0) {
+ dev_err(&tmu->pdev->dev, "failed to remap I/O memory\n");
+ goto err_clk_unprepare;
+ }
- tmu->channels = kzalloc(sizeof(*tmu->channels), GFP_KERNEL);
+ /* Allocate and setup the channels. */
+ if (tmu->model == SH_TMU_LEGACY)
+ tmu->num_channels = 1;
+ else
+ tmu->num_channels = hweight8(cfg->channels_mask);
+
+ tmu->channels = kzalloc(sizeof(*tmu->channels) * tmu->num_channels,
+ GFP_KERNEL);
if (tmu->channels == NULL) {
ret = -ENOMEM;
- goto err3;
+ goto err_unmap;
}
- tmu->num_channels = 1;
-
- tmu->channels[0].base = base;
+ if (tmu->model == SH_TMU_LEGACY) {
+ ret = sh_tmu_channel_setup(&tmu->channels[0], 0,
+ cfg->clockevent_rating != 0,
+ cfg->clocksource_rating != 0, tmu);
+ if (ret < 0)
+ goto err_unmap;
+ } else {
+ /*
+ * Use the first channel as a clock event device and the second
+ * channel as a clock source.
+ */
+ for (i = 0; i < tmu->num_channels; ++i) {
+ ret = sh_tmu_channel_setup(&tmu->channels[i], i,
+ i == 0, i == 1, tmu);
+ if (ret < 0)
+ goto err_unmap;
+ }
+ }
- ret = sh_tmu_channel_setup(&tmu->channels[0], tmu);
- if (ret < 0)
- goto err3;
+ platform_set_drvdata(pdev, tmu);
return 0;
- err3:
+err_unmap:
kfree(tmu->channels);
+ sh_tmu_unmap_memory(tmu);
+err_clk_unprepare:
clk_unprepare(tmu->clk);
- err2:
+err_clk_put:
clk_put(tmu->clk);
- err1:
- iounmap(base);
- err0:
return ret;
}
static int sh_tmu_probe(struct platform_device *pdev)
{
struct sh_tmu_device *tmu = platform_get_drvdata(pdev);
- struct sh_timer_config *cfg = pdev->dev.platform_data;
int ret;
if (!is_early_platform_device(pdev)) {
return 0;
out:
- if (cfg->clockevent_rating || cfg->clocksource_rating)
+ if (tmu->has_clockevent || tmu->has_clocksource)
pm_runtime_irq_safe(&pdev->dev);
else
pm_runtime_idle(&pdev->dev);
return -EBUSY; /* cannot unregister clockevent and clocksource */
}
+static const struct platform_device_id sh_tmu_id_table[] = {
+ { "sh_tmu", SH_TMU_LEGACY },
+ { "sh-tmu", SH_TMU },
+ { "sh-tmu-sh3", SH_TMU_SH3 },
+ { }
+};
+MODULE_DEVICE_TABLE(platform, sh_tmu_id_table);
+
static struct platform_driver sh_tmu_device_driver = {
.probe = sh_tmu_probe,
.remove = sh_tmu_remove,
.driver = {
.name = "sh_tmu",
- }
+ },
+ .id_table = sh_tmu_id_table,
};
static int __init sh_tmu_init(void)