diff options
Diffstat (limited to 'drivers/clocksource')
-rw-r--r-- | drivers/clocksource/Makefile | 2 | ||||
-rw-r--r-- | drivers/clocksource/sh_cmt.c | 116 | ||||
-rw-r--r-- | drivers/clocksource/sh_mtu2.c | 357 | ||||
-rw-r--r-- | drivers/clocksource/sh_tmu.c | 461 |
4 files changed, 917 insertions, 19 deletions
diff --git a/drivers/clocksource/Makefile b/drivers/clocksource/Makefile index 1efb2879a94f..eef216f7f61d 100644 --- a/drivers/clocksource/Makefile +++ b/drivers/clocksource/Makefile | |||
@@ -3,3 +3,5 @@ obj-$(CONFIG_X86_CYCLONE_TIMER) += cyclone.o | |||
3 | obj-$(CONFIG_X86_PM_TIMER) += acpi_pm.o | 3 | obj-$(CONFIG_X86_PM_TIMER) += acpi_pm.o |
4 | obj-$(CONFIG_SCx200HR_TIMER) += scx200_hrt.o | 4 | obj-$(CONFIG_SCx200HR_TIMER) += scx200_hrt.o |
5 | obj-$(CONFIG_SH_TIMER_CMT) += sh_cmt.o | 5 | obj-$(CONFIG_SH_TIMER_CMT) += sh_cmt.o |
6 | obj-$(CONFIG_SH_TIMER_MTU2) += sh_mtu2.o | ||
7 | obj-$(CONFIG_SH_TIMER_TMU) += sh_tmu.o | ||
diff --git a/drivers/clocksource/sh_cmt.c b/drivers/clocksource/sh_cmt.c index 1c92c39a53aa..cf56a2af5fe1 100644 --- a/drivers/clocksource/sh_cmt.c +++ b/drivers/clocksource/sh_cmt.c | |||
@@ -18,7 +18,6 @@ | |||
18 | */ | 18 | */ |
19 | 19 | ||
20 | #include <linux/init.h> | 20 | #include <linux/init.h> |
21 | #include <linux/bootmem.h> | ||
22 | #include <linux/platform_device.h> | 21 | #include <linux/platform_device.h> |
23 | #include <linux/spinlock.h> | 22 | #include <linux/spinlock.h> |
24 | #include <linux/interrupt.h> | 23 | #include <linux/interrupt.h> |
@@ -29,7 +28,7 @@ | |||
29 | #include <linux/err.h> | 28 | #include <linux/err.h> |
30 | #include <linux/clocksource.h> | 29 | #include <linux/clocksource.h> |
31 | #include <linux/clockchips.h> | 30 | #include <linux/clockchips.h> |
32 | #include <linux/sh_cmt.h> | 31 | #include <linux/sh_timer.h> |
33 | 32 | ||
34 | struct sh_cmt_priv { | 33 | struct sh_cmt_priv { |
35 | void __iomem *mapbase; | 34 | void __iomem *mapbase; |
@@ -47,6 +46,7 @@ struct sh_cmt_priv { | |||
47 | unsigned long rate; | 46 | unsigned long rate; |
48 | spinlock_t lock; | 47 | spinlock_t lock; |
49 | struct clock_event_device ced; | 48 | struct clock_event_device ced; |
49 | struct clocksource cs; | ||
50 | unsigned long total_cycles; | 50 | unsigned long total_cycles; |
51 | }; | 51 | }; |
52 | 52 | ||
@@ -59,7 +59,7 @@ static DEFINE_SPINLOCK(sh_cmt_lock); | |||
59 | 59 | ||
60 | static inline unsigned long sh_cmt_read(struct sh_cmt_priv *p, int reg_nr) | 60 | static inline unsigned long sh_cmt_read(struct sh_cmt_priv *p, int reg_nr) |
61 | { | 61 | { |
62 | struct sh_cmt_config *cfg = p->pdev->dev.platform_data; | 62 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; |
63 | void __iomem *base = p->mapbase; | 63 | void __iomem *base = p->mapbase; |
64 | unsigned long offs; | 64 | unsigned long offs; |
65 | 65 | ||
@@ -83,7 +83,7 @@ static inline unsigned long sh_cmt_read(struct sh_cmt_priv *p, int reg_nr) | |||
83 | static inline void sh_cmt_write(struct sh_cmt_priv *p, int reg_nr, | 83 | static inline void sh_cmt_write(struct sh_cmt_priv *p, int reg_nr, |
84 | unsigned long value) | 84 | unsigned long value) |
85 | { | 85 | { |
86 | struct sh_cmt_config *cfg = p->pdev->dev.platform_data; | 86 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; |
87 | void __iomem *base = p->mapbase; | 87 | void __iomem *base = p->mapbase; |
88 | unsigned long offs; | 88 | unsigned long offs; |
89 | 89 | ||
@@ -110,23 +110,28 @@ static unsigned long sh_cmt_get_counter(struct sh_cmt_priv *p, | |||
110 | int *has_wrapped) | 110 | int *has_wrapped) |
111 | { | 111 | { |
112 | unsigned long v1, v2, v3; | 112 | unsigned long v1, v2, v3; |
113 | int o1, o2; | ||
114 | |||
115 | o1 = sh_cmt_read(p, CMCSR) & p->overflow_bit; | ||
113 | 116 | ||
114 | /* Make sure the timer value is stable. Stolen from acpi_pm.c */ | 117 | /* Make sure the timer value is stable. Stolen from acpi_pm.c */ |
115 | do { | 118 | do { |
119 | o2 = o1; | ||
116 | v1 = sh_cmt_read(p, CMCNT); | 120 | v1 = sh_cmt_read(p, CMCNT); |
117 | v2 = sh_cmt_read(p, CMCNT); | 121 | v2 = sh_cmt_read(p, CMCNT); |
118 | v3 = sh_cmt_read(p, CMCNT); | 122 | v3 = sh_cmt_read(p, CMCNT); |
119 | } while (unlikely((v1 > v2 && v1 < v3) || (v2 > v3 && v2 < v1) | 123 | o1 = sh_cmt_read(p, CMCSR) & p->overflow_bit; |
120 | || (v3 > v1 && v3 < v2))); | 124 | } while (unlikely((o1 != o2) || (v1 > v2 && v1 < v3) |
125 | || (v2 > v3 && v2 < v1) || (v3 > v1 && v3 < v2))); | ||
121 | 126 | ||
122 | *has_wrapped = sh_cmt_read(p, CMCSR) & p->overflow_bit; | 127 | *has_wrapped = o1; |
123 | return v2; | 128 | return v2; |
124 | } | 129 | } |
125 | 130 | ||
126 | 131 | ||
127 | static void sh_cmt_start_stop_ch(struct sh_cmt_priv *p, int start) | 132 | static void sh_cmt_start_stop_ch(struct sh_cmt_priv *p, int start) |
128 | { | 133 | { |
129 | struct sh_cmt_config *cfg = p->pdev->dev.platform_data; | 134 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; |
130 | unsigned long flags, value; | 135 | unsigned long flags, value; |
131 | 136 | ||
132 | /* start stop register shared by multiple timer channels */ | 137 | /* start stop register shared by multiple timer channels */ |
@@ -144,7 +149,7 @@ static void sh_cmt_start_stop_ch(struct sh_cmt_priv *p, int start) | |||
144 | 149 | ||
145 | static int sh_cmt_enable(struct sh_cmt_priv *p, unsigned long *rate) | 150 | static int sh_cmt_enable(struct sh_cmt_priv *p, unsigned long *rate) |
146 | { | 151 | { |
147 | struct sh_cmt_config *cfg = p->pdev->dev.platform_data; | 152 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; |
148 | int ret; | 153 | int ret; |
149 | 154 | ||
150 | /* enable clock */ | 155 | /* enable clock */ |
@@ -153,16 +158,18 @@ static int sh_cmt_enable(struct sh_cmt_priv *p, unsigned long *rate) | |||
153 | pr_err("sh_cmt: cannot enable clock \"%s\"\n", cfg->clk); | 158 | pr_err("sh_cmt: cannot enable clock \"%s\"\n", cfg->clk); |
154 | return ret; | 159 | return ret; |
155 | } | 160 | } |
156 | *rate = clk_get_rate(p->clk) / 8; | ||
157 | 161 | ||
158 | /* make sure channel is disabled */ | 162 | /* make sure channel is disabled */ |
159 | sh_cmt_start_stop_ch(p, 0); | 163 | sh_cmt_start_stop_ch(p, 0); |
160 | 164 | ||
161 | /* configure channel, periodic mode and maximum timeout */ | 165 | /* configure channel, periodic mode and maximum timeout */ |
162 | if (p->width == 16) | 166 | if (p->width == 16) { |
163 | sh_cmt_write(p, CMCSR, 0); | 167 | *rate = clk_get_rate(p->clk) / 512; |
164 | else | 168 | sh_cmt_write(p, CMCSR, 0x43); |
169 | } else { | ||
170 | *rate = clk_get_rate(p->clk) / 8; | ||
165 | sh_cmt_write(p, CMCSR, 0x01a4); | 171 | sh_cmt_write(p, CMCSR, 0x01a4); |
172 | } | ||
166 | 173 | ||
167 | sh_cmt_write(p, CMCOR, 0xffffffff); | 174 | sh_cmt_write(p, CMCOR, 0xffffffff); |
168 | sh_cmt_write(p, CMCNT, 0); | 175 | sh_cmt_write(p, CMCNT, 0); |
@@ -376,6 +383,68 @@ static void sh_cmt_stop(struct sh_cmt_priv *p, unsigned long flag) | |||
376 | spin_unlock_irqrestore(&p->lock, flags); | 383 | spin_unlock_irqrestore(&p->lock, flags); |
377 | } | 384 | } |
378 | 385 | ||
386 | static struct sh_cmt_priv *cs_to_sh_cmt(struct clocksource *cs) | ||
387 | { | ||
388 | return container_of(cs, struct sh_cmt_priv, cs); | ||
389 | } | ||
390 | |||
391 | static cycle_t sh_cmt_clocksource_read(struct clocksource *cs) | ||
392 | { | ||
393 | struct sh_cmt_priv *p = cs_to_sh_cmt(cs); | ||
394 | unsigned long flags, raw; | ||
395 | unsigned long value; | ||
396 | int has_wrapped; | ||
397 | |||
398 | spin_lock_irqsave(&p->lock, flags); | ||
399 | value = p->total_cycles; | ||
400 | raw = sh_cmt_get_counter(p, &has_wrapped); | ||
401 | |||
402 | if (unlikely(has_wrapped)) | ||
403 | raw += p->match_value; | ||
404 | spin_unlock_irqrestore(&p->lock, flags); | ||
405 | |||
406 | return value + raw; | ||
407 | } | ||
408 | |||
409 | static int sh_cmt_clocksource_enable(struct clocksource *cs) | ||
410 | { | ||
411 | struct sh_cmt_priv *p = cs_to_sh_cmt(cs); | ||
412 | int ret; | ||
413 | |||
414 | p->total_cycles = 0; | ||
415 | |||
416 | ret = sh_cmt_start(p, FLAG_CLOCKSOURCE); | ||
417 | if (ret) | ||
418 | return ret; | ||
419 | |||
420 | /* TODO: calculate good shift from rate and counter bit width */ | ||
421 | cs->shift = 0; | ||
422 | cs->mult = clocksource_hz2mult(p->rate, cs->shift); | ||
423 | return 0; | ||
424 | } | ||
425 | |||
426 | static void sh_cmt_clocksource_disable(struct clocksource *cs) | ||
427 | { | ||
428 | sh_cmt_stop(cs_to_sh_cmt(cs), FLAG_CLOCKSOURCE); | ||
429 | } | ||
430 | |||
431 | static int sh_cmt_register_clocksource(struct sh_cmt_priv *p, | ||
432 | char *name, unsigned long rating) | ||
433 | { | ||
434 | struct clocksource *cs = &p->cs; | ||
435 | |||
436 | cs->name = name; | ||
437 | cs->rating = rating; | ||
438 | cs->read = sh_cmt_clocksource_read; | ||
439 | cs->enable = sh_cmt_clocksource_enable; | ||
440 | cs->disable = sh_cmt_clocksource_disable; | ||
441 | cs->mask = CLOCKSOURCE_MASK(sizeof(unsigned long) * 8); | ||
442 | cs->flags = CLOCK_SOURCE_IS_CONTINUOUS; | ||
443 | pr_info("sh_cmt: %s used as clock source\n", cs->name); | ||
444 | clocksource_register(cs); | ||
445 | return 0; | ||
446 | } | ||
447 | |||
379 | static struct sh_cmt_priv *ced_to_sh_cmt(struct clock_event_device *ced) | 448 | static struct sh_cmt_priv *ced_to_sh_cmt(struct clock_event_device *ced) |
380 | { | 449 | { |
381 | return container_of(ced, struct sh_cmt_priv, ced); | 450 | return container_of(ced, struct sh_cmt_priv, ced); |
@@ -468,9 +537,9 @@ static void sh_cmt_register_clockevent(struct sh_cmt_priv *p, | |||
468 | clockevents_register_device(ced); | 537 | clockevents_register_device(ced); |
469 | } | 538 | } |
470 | 539 | ||
471 | int sh_cmt_register(struct sh_cmt_priv *p, char *name, | 540 | static int sh_cmt_register(struct sh_cmt_priv *p, char *name, |
472 | unsigned long clockevent_rating, | 541 | unsigned long clockevent_rating, |
473 | unsigned long clocksource_rating) | 542 | unsigned long clocksource_rating) |
474 | { | 543 | { |
475 | if (p->width == (sizeof(p->max_match_value) * 8)) | 544 | if (p->width == (sizeof(p->max_match_value) * 8)) |
476 | p->max_match_value = ~0; | 545 | p->max_match_value = ~0; |
@@ -483,12 +552,15 @@ int sh_cmt_register(struct sh_cmt_priv *p, char *name, | |||
483 | if (clockevent_rating) | 552 | if (clockevent_rating) |
484 | sh_cmt_register_clockevent(p, name, clockevent_rating); | 553 | sh_cmt_register_clockevent(p, name, clockevent_rating); |
485 | 554 | ||
555 | if (clocksource_rating) | ||
556 | sh_cmt_register_clocksource(p, name, clocksource_rating); | ||
557 | |||
486 | return 0; | 558 | return 0; |
487 | } | 559 | } |
488 | 560 | ||
489 | static int sh_cmt_setup(struct sh_cmt_priv *p, struct platform_device *pdev) | 561 | static int sh_cmt_setup(struct sh_cmt_priv *p, struct platform_device *pdev) |
490 | { | 562 | { |
491 | struct sh_cmt_config *cfg = pdev->dev.platform_data; | 563 | struct sh_timer_config *cfg = pdev->dev.platform_data; |
492 | struct resource *res; | 564 | struct resource *res; |
493 | int irq, ret; | 565 | int irq, ret; |
494 | ret = -ENXIO; | 566 | ret = -ENXIO; |
@@ -545,7 +617,7 @@ static int sh_cmt_setup(struct sh_cmt_priv *p, struct platform_device *pdev) | |||
545 | if (resource_size(res) == 6) { | 617 | if (resource_size(res) == 6) { |
546 | p->width = 16; | 618 | p->width = 16; |
547 | p->overflow_bit = 0x80; | 619 | p->overflow_bit = 0x80; |
548 | p->clear_bits = ~0xc0; | 620 | p->clear_bits = ~0x80; |
549 | } else { | 621 | } else { |
550 | p->width = 32; | 622 | p->width = 32; |
551 | p->overflow_bit = 0x8000; | 623 | p->overflow_bit = 0x8000; |
@@ -566,8 +638,14 @@ static int sh_cmt_setup(struct sh_cmt_priv *p, struct platform_device *pdev) | |||
566 | static int __devinit sh_cmt_probe(struct platform_device *pdev) | 638 | static int __devinit sh_cmt_probe(struct platform_device *pdev) |
567 | { | 639 | { |
568 | struct sh_cmt_priv *p = platform_get_drvdata(pdev); | 640 | struct sh_cmt_priv *p = platform_get_drvdata(pdev); |
641 | struct sh_timer_config *cfg = pdev->dev.platform_data; | ||
569 | int ret; | 642 | int ret; |
570 | 643 | ||
644 | if (p) { | ||
645 | pr_info("sh_cmt: %s kept as earlytimer\n", cfg->name); | ||
646 | return 0; | ||
647 | } | ||
648 | |||
571 | p = kmalloc(sizeof(*p), GFP_KERNEL); | 649 | p = kmalloc(sizeof(*p), GFP_KERNEL); |
572 | if (p == NULL) { | 650 | if (p == NULL) { |
573 | dev_err(&pdev->dev, "failed to allocate driver data\n"); | 651 | dev_err(&pdev->dev, "failed to allocate driver data\n"); |
@@ -577,7 +655,6 @@ static int __devinit sh_cmt_probe(struct platform_device *pdev) | |||
577 | ret = sh_cmt_setup(p, pdev); | 655 | ret = sh_cmt_setup(p, pdev); |
578 | if (ret) { | 656 | if (ret) { |
579 | kfree(p); | 657 | kfree(p); |
580 | |||
581 | platform_set_drvdata(pdev, NULL); | 658 | platform_set_drvdata(pdev, NULL); |
582 | } | 659 | } |
583 | return ret; | 660 | return ret; |
@@ -606,6 +683,7 @@ static void __exit sh_cmt_exit(void) | |||
606 | platform_driver_unregister(&sh_cmt_device_driver); | 683 | platform_driver_unregister(&sh_cmt_device_driver); |
607 | } | 684 | } |
608 | 685 | ||
686 | early_platform_init("earlytimer", &sh_cmt_device_driver); | ||
609 | module_init(sh_cmt_init); | 687 | module_init(sh_cmt_init); |
610 | module_exit(sh_cmt_exit); | 688 | module_exit(sh_cmt_exit); |
611 | 689 | ||
diff --git a/drivers/clocksource/sh_mtu2.c b/drivers/clocksource/sh_mtu2.c new file mode 100644 index 000000000000..d1ae75454d10 --- /dev/null +++ b/drivers/clocksource/sh_mtu2.c | |||
@@ -0,0 +1,357 @@ | |||
1 | /* | ||
2 | * SuperH Timer Support - MTU2 | ||
3 | * | ||
4 | * Copyright (C) 2009 Magnus Damm | ||
5 | * | ||
6 | * This program is free software; you can redistribute it and/or modify | ||
7 | * it under the terms of the GNU General Public License as published by | ||
8 | * the Free Software Foundation; either version 2 of the License | ||
9 | * | ||
10 | * This program is distributed in the hope that it will be useful, | ||
11 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
12 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
13 | * GNU General Public License for more details. | ||
14 | * | ||
15 | * You should have received a copy of the GNU General Public License | ||
16 | * along with this program; if not, write to the Free Software | ||
17 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA | ||
18 | */ | ||
19 | |||
20 | #include <linux/init.h> | ||
21 | #include <linux/platform_device.h> | ||
22 | #include <linux/spinlock.h> | ||
23 | #include <linux/interrupt.h> | ||
24 | #include <linux/ioport.h> | ||
25 | #include <linux/delay.h> | ||
26 | #include <linux/io.h> | ||
27 | #include <linux/clk.h> | ||
28 | #include <linux/irq.h> | ||
29 | #include <linux/err.h> | ||
30 | #include <linux/clockchips.h> | ||
31 | #include <linux/sh_timer.h> | ||
32 | |||
33 | struct sh_mtu2_priv { | ||
34 | void __iomem *mapbase; | ||
35 | struct clk *clk; | ||
36 | struct irqaction irqaction; | ||
37 | struct platform_device *pdev; | ||
38 | unsigned long rate; | ||
39 | unsigned long periodic; | ||
40 | struct clock_event_device ced; | ||
41 | }; | ||
42 | |||
43 | static DEFINE_SPINLOCK(sh_mtu2_lock); | ||
44 | |||
45 | #define TSTR -1 /* shared register */ | ||
46 | #define TCR 0 /* channel register */ | ||
47 | #define TMDR 1 /* channel register */ | ||
48 | #define TIOR 2 /* channel register */ | ||
49 | #define TIER 3 /* channel register */ | ||
50 | #define TSR 4 /* channel register */ | ||
51 | #define TCNT 5 /* channel register */ | ||
52 | #define TGR 6 /* channel register */ | ||
53 | |||
54 | static unsigned long mtu2_reg_offs[] = { | ||
55 | [TCR] = 0, | ||
56 | [TMDR] = 1, | ||
57 | [TIOR] = 2, | ||
58 | [TIER] = 4, | ||
59 | [TSR] = 5, | ||
60 | [TCNT] = 6, | ||
61 | [TGR] = 8, | ||
62 | }; | ||
63 | |||
64 | static inline unsigned long sh_mtu2_read(struct sh_mtu2_priv *p, int reg_nr) | ||
65 | { | ||
66 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; | ||
67 | void __iomem *base = p->mapbase; | ||
68 | unsigned long offs; | ||
69 | |||
70 | if (reg_nr == TSTR) | ||
71 | return ioread8(base + cfg->channel_offset); | ||
72 | |||
73 | offs = mtu2_reg_offs[reg_nr]; | ||
74 | |||
75 | if ((reg_nr == TCNT) || (reg_nr == TGR)) | ||
76 | return ioread16(base + offs); | ||
77 | else | ||
78 | return ioread8(base + offs); | ||
79 | } | ||
80 | |||
81 | static inline void sh_mtu2_write(struct sh_mtu2_priv *p, int reg_nr, | ||
82 | unsigned long value) | ||
83 | { | ||
84 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; | ||
85 | void __iomem *base = p->mapbase; | ||
86 | unsigned long offs; | ||
87 | |||
88 | if (reg_nr == TSTR) { | ||
89 | iowrite8(value, base + cfg->channel_offset); | ||
90 | return; | ||
91 | } | ||
92 | |||
93 | offs = mtu2_reg_offs[reg_nr]; | ||
94 | |||
95 | if ((reg_nr == TCNT) || (reg_nr == TGR)) | ||
96 | iowrite16(value, base + offs); | ||
97 | else | ||
98 | iowrite8(value, base + offs); | ||
99 | } | ||
100 | |||
101 | static void sh_mtu2_start_stop_ch(struct sh_mtu2_priv *p, int start) | ||
102 | { | ||
103 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; | ||
104 | unsigned long flags, value; | ||
105 | |||
106 | /* start stop register shared by multiple timer channels */ | ||
107 | spin_lock_irqsave(&sh_mtu2_lock, flags); | ||
108 | value = sh_mtu2_read(p, TSTR); | ||
109 | |||
110 | if (start) | ||
111 | value |= 1 << cfg->timer_bit; | ||
112 | else | ||
113 | value &= ~(1 << cfg->timer_bit); | ||
114 | |||
115 | sh_mtu2_write(p, TSTR, value); | ||
116 | spin_unlock_irqrestore(&sh_mtu2_lock, flags); | ||
117 | } | ||
118 | |||
119 | static int sh_mtu2_enable(struct sh_mtu2_priv *p) | ||
120 | { | ||
121 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; | ||
122 | int ret; | ||
123 | |||
124 | /* enable clock */ | ||
125 | ret = clk_enable(p->clk); | ||
126 | if (ret) { | ||
127 | pr_err("sh_mtu2: cannot enable clock \"%s\"\n", cfg->clk); | ||
128 | return ret; | ||
129 | } | ||
130 | |||
131 | /* make sure channel is disabled */ | ||
132 | sh_mtu2_start_stop_ch(p, 0); | ||
133 | |||
134 | p->rate = clk_get_rate(p->clk) / 64; | ||
135 | p->periodic = (p->rate + HZ/2) / HZ; | ||
136 | |||
137 | /* "Periodic Counter Operation" */ | ||
138 | sh_mtu2_write(p, TCR, 0x23); /* TGRA clear, divide clock by 64 */ | ||
139 | sh_mtu2_write(p, TIOR, 0); | ||
140 | sh_mtu2_write(p, TGR, p->periodic); | ||
141 | sh_mtu2_write(p, TCNT, 0); | ||
142 | sh_mtu2_write(p, TMDR, 0); | ||
143 | sh_mtu2_write(p, TIER, 0x01); | ||
144 | |||
145 | /* enable channel */ | ||
146 | sh_mtu2_start_stop_ch(p, 1); | ||
147 | |||
148 | return 0; | ||
149 | } | ||
150 | |||
151 | static void sh_mtu2_disable(struct sh_mtu2_priv *p) | ||
152 | { | ||
153 | /* disable channel */ | ||
154 | sh_mtu2_start_stop_ch(p, 0); | ||
155 | |||
156 | /* stop clock */ | ||
157 | clk_disable(p->clk); | ||
158 | } | ||
159 | |||
160 | static irqreturn_t sh_mtu2_interrupt(int irq, void *dev_id) | ||
161 | { | ||
162 | struct sh_mtu2_priv *p = dev_id; | ||
163 | |||
164 | /* acknowledge interrupt */ | ||
165 | sh_mtu2_read(p, TSR); | ||
166 | sh_mtu2_write(p, TSR, 0xfe); | ||
167 | |||
168 | /* notify clockevent layer */ | ||
169 | p->ced.event_handler(&p->ced); | ||
170 | return IRQ_HANDLED; | ||
171 | } | ||
172 | |||
173 | static struct sh_mtu2_priv *ced_to_sh_mtu2(struct clock_event_device *ced) | ||
174 | { | ||
175 | return container_of(ced, struct sh_mtu2_priv, ced); | ||
176 | } | ||
177 | |||
178 | static void sh_mtu2_clock_event_mode(enum clock_event_mode mode, | ||
179 | struct clock_event_device *ced) | ||
180 | { | ||
181 | struct sh_mtu2_priv *p = ced_to_sh_mtu2(ced); | ||
182 | int disabled = 0; | ||
183 | |||
184 | /* deal with old setting first */ | ||
185 | switch (ced->mode) { | ||
186 | case CLOCK_EVT_MODE_PERIODIC: | ||
187 | sh_mtu2_disable(p); | ||
188 | disabled = 1; | ||
189 | break; | ||
190 | default: | ||
191 | break; | ||
192 | } | ||
193 | |||
194 | switch (mode) { | ||
195 | case CLOCK_EVT_MODE_PERIODIC: | ||
196 | pr_info("sh_mtu2: %s used for periodic clock events\n", | ||
197 | ced->name); | ||
198 | sh_mtu2_enable(p); | ||
199 | break; | ||
200 | case CLOCK_EVT_MODE_UNUSED: | ||
201 | if (!disabled) | ||
202 | sh_mtu2_disable(p); | ||
203 | break; | ||
204 | case CLOCK_EVT_MODE_SHUTDOWN: | ||
205 | default: | ||
206 | break; | ||
207 | } | ||
208 | } | ||
209 | |||
210 | static void sh_mtu2_register_clockevent(struct sh_mtu2_priv *p, | ||
211 | char *name, unsigned long rating) | ||
212 | { | ||
213 | struct clock_event_device *ced = &p->ced; | ||
214 | int ret; | ||
215 | |||
216 | memset(ced, 0, sizeof(*ced)); | ||
217 | |||
218 | ced->name = name; | ||
219 | ced->features = CLOCK_EVT_FEAT_PERIODIC; | ||
220 | ced->rating = rating; | ||
221 | ced->cpumask = cpumask_of(0); | ||
222 | ced->set_mode = sh_mtu2_clock_event_mode; | ||
223 | |||
224 | ret = setup_irq(p->irqaction.irq, &p->irqaction); | ||
225 | if (ret) { | ||
226 | pr_err("sh_mtu2: failed to request irq %d\n", | ||
227 | p->irqaction.irq); | ||
228 | return; | ||
229 | } | ||
230 | |||
231 | pr_info("sh_mtu2: %s used for clock events\n", ced->name); | ||
232 | clockevents_register_device(ced); | ||
233 | } | ||
234 | |||
235 | static int sh_mtu2_register(struct sh_mtu2_priv *p, char *name, | ||
236 | unsigned long clockevent_rating) | ||
237 | { | ||
238 | if (clockevent_rating) | ||
239 | sh_mtu2_register_clockevent(p, name, clockevent_rating); | ||
240 | |||
241 | return 0; | ||
242 | } | ||
243 | |||
244 | static int sh_mtu2_setup(struct sh_mtu2_priv *p, struct platform_device *pdev) | ||
245 | { | ||
246 | struct sh_timer_config *cfg = pdev->dev.platform_data; | ||
247 | struct resource *res; | ||
248 | int irq, ret; | ||
249 | ret = -ENXIO; | ||
250 | |||
251 | memset(p, 0, sizeof(*p)); | ||
252 | p->pdev = pdev; | ||
253 | |||
254 | if (!cfg) { | ||
255 | dev_err(&p->pdev->dev, "missing platform data\n"); | ||
256 | goto err0; | ||
257 | } | ||
258 | |||
259 | platform_set_drvdata(pdev, p); | ||
260 | |||
261 | res = platform_get_resource(p->pdev, IORESOURCE_MEM, 0); | ||
262 | if (!res) { | ||
263 | dev_err(&p->pdev->dev, "failed to get I/O memory\n"); | ||
264 | goto err0; | ||
265 | } | ||
266 | |||
267 | irq = platform_get_irq(p->pdev, 0); | ||
268 | if (irq < 0) { | ||
269 | dev_err(&p->pdev->dev, "failed to get irq\n"); | ||
270 | goto err0; | ||
271 | } | ||
272 | |||
273 | /* map memory, let mapbase point to our channel */ | ||
274 | p->mapbase = ioremap_nocache(res->start, resource_size(res)); | ||
275 | if (p->mapbase == NULL) { | ||
276 | pr_err("sh_mtu2: failed to remap I/O memory\n"); | ||
277 | goto err0; | ||
278 | } | ||
279 | |||
280 | /* setup data for setup_irq() (too early for request_irq()) */ | ||
281 | p->irqaction.name = cfg->name; | ||
282 | p->irqaction.handler = sh_mtu2_interrupt; | ||
283 | p->irqaction.dev_id = p; | ||
284 | p->irqaction.irq = irq; | ||
285 | p->irqaction.flags = IRQF_DISABLED | IRQF_TIMER | IRQF_IRQPOLL; | ||
286 | p->irqaction.mask = CPU_MASK_NONE; | ||
287 | |||
288 | /* get hold of clock */ | ||
289 | p->clk = clk_get(&p->pdev->dev, cfg->clk); | ||
290 | if (IS_ERR(p->clk)) { | ||
291 | pr_err("sh_mtu2: cannot get clock \"%s\"\n", cfg->clk); | ||
292 | ret = PTR_ERR(p->clk); | ||
293 | goto err1; | ||
294 | } | ||
295 | |||
296 | return sh_mtu2_register(p, cfg->name, cfg->clockevent_rating); | ||
297 | err1: | ||
298 | iounmap(p->mapbase); | ||
299 | err0: | ||
300 | return ret; | ||
301 | } | ||
302 | |||
303 | static int __devinit sh_mtu2_probe(struct platform_device *pdev) | ||
304 | { | ||
305 | struct sh_mtu2_priv *p = platform_get_drvdata(pdev); | ||
306 | struct sh_timer_config *cfg = pdev->dev.platform_data; | ||
307 | int ret; | ||
308 | |||
309 | if (p) { | ||
310 | pr_info("sh_mtu2: %s kept as earlytimer\n", cfg->name); | ||
311 | return 0; | ||
312 | } | ||
313 | |||
314 | p = kmalloc(sizeof(*p), GFP_KERNEL); | ||
315 | if (p == NULL) { | ||
316 | dev_err(&pdev->dev, "failed to allocate driver data\n"); | ||
317 | return -ENOMEM; | ||
318 | } | ||
319 | |||
320 | ret = sh_mtu2_setup(p, pdev); | ||
321 | if (ret) { | ||
322 | kfree(p); | ||
323 | platform_set_drvdata(pdev, NULL); | ||
324 | } | ||
325 | return ret; | ||
326 | } | ||
327 | |||
328 | static int __devexit sh_mtu2_remove(struct platform_device *pdev) | ||
329 | { | ||
330 | return -EBUSY; /* cannot unregister clockevent */ | ||
331 | } | ||
332 | |||
333 | static struct platform_driver sh_mtu2_device_driver = { | ||
334 | .probe = sh_mtu2_probe, | ||
335 | .remove = __devexit_p(sh_mtu2_remove), | ||
336 | .driver = { | ||
337 | .name = "sh_mtu2", | ||
338 | } | ||
339 | }; | ||
340 | |||
341 | static int __init sh_mtu2_init(void) | ||
342 | { | ||
343 | return platform_driver_register(&sh_mtu2_device_driver); | ||
344 | } | ||
345 | |||
346 | static void __exit sh_mtu2_exit(void) | ||
347 | { | ||
348 | platform_driver_unregister(&sh_mtu2_device_driver); | ||
349 | } | ||
350 | |||
351 | early_platform_init("earlytimer", &sh_mtu2_device_driver); | ||
352 | module_init(sh_mtu2_init); | ||
353 | module_exit(sh_mtu2_exit); | ||
354 | |||
355 | MODULE_AUTHOR("Magnus Damm"); | ||
356 | MODULE_DESCRIPTION("SuperH MTU2 Timer Driver"); | ||
357 | MODULE_LICENSE("GPL v2"); | ||
diff --git a/drivers/clocksource/sh_tmu.c b/drivers/clocksource/sh_tmu.c new file mode 100644 index 000000000000..d6ea4398bf62 --- /dev/null +++ b/drivers/clocksource/sh_tmu.c | |||
@@ -0,0 +1,461 @@ | |||
1 | /* | ||
2 | * SuperH Timer Support - TMU | ||
3 | * | ||
4 | * Copyright (C) 2009 Magnus Damm | ||
5 | * | ||
6 | * This program is free software; you can redistribute it and/or modify | ||
7 | * it under the terms of the GNU General Public License as published by | ||
8 | * the Free Software Foundation; either version 2 of the License | ||
9 | * | ||
10 | * This program is distributed in the hope that it will be useful, | ||
11 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
12 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
13 | * GNU General Public License for more details. | ||
14 | * | ||
15 | * You should have received a copy of the GNU General Public License | ||
16 | * along with this program; if not, write to the Free Software | ||
17 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA | ||
18 | */ | ||
19 | |||
20 | #include <linux/init.h> | ||
21 | #include <linux/platform_device.h> | ||
22 | #include <linux/spinlock.h> | ||
23 | #include <linux/interrupt.h> | ||
24 | #include <linux/ioport.h> | ||
25 | #include <linux/delay.h> | ||
26 | #include <linux/io.h> | ||
27 | #include <linux/clk.h> | ||
28 | #include <linux/irq.h> | ||
29 | #include <linux/err.h> | ||
30 | #include <linux/clocksource.h> | ||
31 | #include <linux/clockchips.h> | ||
32 | #include <linux/sh_timer.h> | ||
33 | |||
34 | struct sh_tmu_priv { | ||
35 | void __iomem *mapbase; | ||
36 | struct clk *clk; | ||
37 | struct irqaction irqaction; | ||
38 | struct platform_device *pdev; | ||
39 | unsigned long rate; | ||
40 | unsigned long periodic; | ||
41 | struct clock_event_device ced; | ||
42 | struct clocksource cs; | ||
43 | }; | ||
44 | |||
45 | static DEFINE_SPINLOCK(sh_tmu_lock); | ||
46 | |||
47 | #define TSTR -1 /* shared register */ | ||
48 | #define TCOR 0 /* channel register */ | ||
49 | #define TCNT 1 /* channel register */ | ||
50 | #define TCR 2 /* channel register */ | ||
51 | |||
52 | static inline unsigned long sh_tmu_read(struct sh_tmu_priv *p, int reg_nr) | ||
53 | { | ||
54 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; | ||
55 | void __iomem *base = p->mapbase; | ||
56 | unsigned long offs; | ||
57 | |||
58 | if (reg_nr == TSTR) | ||
59 | return ioread8(base - cfg->channel_offset); | ||
60 | |||
61 | offs = reg_nr << 2; | ||
62 | |||
63 | if (reg_nr == TCR) | ||
64 | return ioread16(base + offs); | ||
65 | else | ||
66 | return ioread32(base + offs); | ||
67 | } | ||
68 | |||
69 | static inline void sh_tmu_write(struct sh_tmu_priv *p, int reg_nr, | ||
70 | unsigned long value) | ||
71 | { | ||
72 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; | ||
73 | void __iomem *base = p->mapbase; | ||
74 | unsigned long offs; | ||
75 | |||
76 | if (reg_nr == TSTR) { | ||
77 | iowrite8(value, base - cfg->channel_offset); | ||
78 | return; | ||
79 | } | ||
80 | |||
81 | offs = reg_nr << 2; | ||
82 | |||
83 | if (reg_nr == TCR) | ||
84 | iowrite16(value, base + offs); | ||
85 | else | ||
86 | iowrite32(value, base + offs); | ||
87 | } | ||
88 | |||
89 | static void sh_tmu_start_stop_ch(struct sh_tmu_priv *p, int start) | ||
90 | { | ||
91 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; | ||
92 | unsigned long flags, value; | ||
93 | |||
94 | /* start stop register shared by multiple timer channels */ | ||
95 | spin_lock_irqsave(&sh_tmu_lock, flags); | ||
96 | value = sh_tmu_read(p, TSTR); | ||
97 | |||
98 | if (start) | ||
99 | value |= 1 << cfg->timer_bit; | ||
100 | else | ||
101 | value &= ~(1 << cfg->timer_bit); | ||
102 | |||
103 | sh_tmu_write(p, TSTR, value); | ||
104 | spin_unlock_irqrestore(&sh_tmu_lock, flags); | ||
105 | } | ||
106 | |||
107 | static int sh_tmu_enable(struct sh_tmu_priv *p) | ||
108 | { | ||
109 | struct sh_timer_config *cfg = p->pdev->dev.platform_data; | ||
110 | int ret; | ||
111 | |||
112 | /* enable clock */ | ||
113 | ret = clk_enable(p->clk); | ||
114 | if (ret) { | ||
115 | pr_err("sh_tmu: cannot enable clock \"%s\"\n", cfg->clk); | ||
116 | return ret; | ||
117 | } | ||
118 | |||
119 | /* make sure channel is disabled */ | ||
120 | sh_tmu_start_stop_ch(p, 0); | ||
121 | |||
122 | /* maximum timeout */ | ||
123 | sh_tmu_write(p, TCOR, 0xffffffff); | ||
124 | sh_tmu_write(p, TCNT, 0xffffffff); | ||
125 | |||
126 | /* configure channel to parent clock / 4, irq off */ | ||
127 | p->rate = clk_get_rate(p->clk) / 4; | ||
128 | sh_tmu_write(p, TCR, 0x0000); | ||
129 | |||
130 | /* enable channel */ | ||
131 | sh_tmu_start_stop_ch(p, 1); | ||
132 | |||
133 | return 0; | ||
134 | } | ||
135 | |||
136 | static void sh_tmu_disable(struct sh_tmu_priv *p) | ||
137 | { | ||
138 | /* disable channel */ | ||
139 | sh_tmu_start_stop_ch(p, 0); | ||
140 | |||
141 | /* stop clock */ | ||
142 | clk_disable(p->clk); | ||
143 | } | ||
144 | |||
145 | static void sh_tmu_set_next(struct sh_tmu_priv *p, unsigned long delta, | ||
146 | int periodic) | ||
147 | { | ||
148 | /* stop timer */ | ||
149 | sh_tmu_start_stop_ch(p, 0); | ||
150 | |||
151 | /* acknowledge interrupt */ | ||
152 | sh_tmu_read(p, TCR); | ||
153 | |||
154 | /* enable interrupt */ | ||
155 | sh_tmu_write(p, TCR, 0x0020); | ||
156 | |||
157 | /* reload delta value in case of periodic timer */ | ||
158 | if (periodic) | ||
159 | sh_tmu_write(p, TCOR, delta); | ||
160 | else | ||
161 | sh_tmu_write(p, TCOR, 0); | ||
162 | |||
163 | sh_tmu_write(p, TCNT, delta); | ||
164 | |||
165 | /* start timer */ | ||
166 | sh_tmu_start_stop_ch(p, 1); | ||
167 | } | ||
168 | |||
169 | static irqreturn_t sh_tmu_interrupt(int irq, void *dev_id) | ||
170 | { | ||
171 | struct sh_tmu_priv *p = dev_id; | ||
172 | |||
173 | /* disable or acknowledge interrupt */ | ||
174 | if (p->ced.mode == CLOCK_EVT_MODE_ONESHOT) | ||
175 | sh_tmu_write(p, TCR, 0x0000); | ||
176 | else | ||
177 | sh_tmu_write(p, TCR, 0x0020); | ||
178 | |||
179 | /* notify clockevent layer */ | ||
180 | p->ced.event_handler(&p->ced); | ||
181 | return IRQ_HANDLED; | ||
182 | } | ||
183 | |||
184 | static struct sh_tmu_priv *cs_to_sh_tmu(struct clocksource *cs) | ||
185 | { | ||
186 | return container_of(cs, struct sh_tmu_priv, cs); | ||
187 | } | ||
188 | |||
189 | static cycle_t sh_tmu_clocksource_read(struct clocksource *cs) | ||
190 | { | ||
191 | struct sh_tmu_priv *p = cs_to_sh_tmu(cs); | ||
192 | |||
193 | return sh_tmu_read(p, TCNT) ^ 0xffffffff; | ||
194 | } | ||
195 | |||
196 | static int sh_tmu_clocksource_enable(struct clocksource *cs) | ||
197 | { | ||
198 | struct sh_tmu_priv *p = cs_to_sh_tmu(cs); | ||
199 | int ret; | ||
200 | |||
201 | ret = sh_tmu_enable(p); | ||
202 | if (ret) | ||
203 | return ret; | ||
204 | |||
205 | /* TODO: calculate good shift from rate and counter bit width */ | ||
206 | cs->shift = 10; | ||
207 | cs->mult = clocksource_hz2mult(p->rate, cs->shift); | ||
208 | return 0; | ||
209 | } | ||
210 | |||
211 | static void sh_tmu_clocksource_disable(struct clocksource *cs) | ||
212 | { | ||
213 | sh_tmu_disable(cs_to_sh_tmu(cs)); | ||
214 | } | ||
215 | |||
216 | static int sh_tmu_register_clocksource(struct sh_tmu_priv *p, | ||
217 | char *name, unsigned long rating) | ||
218 | { | ||
219 | struct clocksource *cs = &p->cs; | ||
220 | |||
221 | cs->name = name; | ||
222 | cs->rating = rating; | ||
223 | cs->read = sh_tmu_clocksource_read; | ||
224 | cs->enable = sh_tmu_clocksource_enable; | ||
225 | cs->disable = sh_tmu_clocksource_disable; | ||
226 | cs->mask = CLOCKSOURCE_MASK(32); | ||
227 | cs->flags = CLOCK_SOURCE_IS_CONTINUOUS; | ||
228 | pr_info("sh_tmu: %s used as clock source\n", cs->name); | ||
229 | clocksource_register(cs); | ||
230 | return 0; | ||
231 | } | ||
232 | |||
233 | static struct sh_tmu_priv *ced_to_sh_tmu(struct clock_event_device *ced) | ||
234 | { | ||
235 | return container_of(ced, struct sh_tmu_priv, ced); | ||
236 | } | ||
237 | |||
238 | static void sh_tmu_clock_event_start(struct sh_tmu_priv *p, int periodic) | ||
239 | { | ||
240 | struct clock_event_device *ced = &p->ced; | ||
241 | |||
242 | sh_tmu_enable(p); | ||
243 | |||
244 | /* TODO: calculate good shift from rate and counter bit width */ | ||
245 | |||
246 | ced->shift = 32; | ||
247 | ced->mult = div_sc(p->rate, NSEC_PER_SEC, ced->shift); | ||
248 | ced->max_delta_ns = clockevent_delta2ns(0xffffffff, ced); | ||
249 | ced->min_delta_ns = 5000; | ||
250 | |||
251 | if (periodic) { | ||
252 | p->periodic = (p->rate + HZ/2) / HZ; | ||
253 | sh_tmu_set_next(p, p->periodic, 1); | ||
254 | } | ||
255 | } | ||
256 | |||
257 | static void sh_tmu_clock_event_mode(enum clock_event_mode mode, | ||
258 | struct clock_event_device *ced) | ||
259 | { | ||
260 | struct sh_tmu_priv *p = ced_to_sh_tmu(ced); | ||
261 | int disabled = 0; | ||
262 | |||
263 | /* deal with old setting first */ | ||
264 | switch (ced->mode) { | ||
265 | case CLOCK_EVT_MODE_PERIODIC: | ||
266 | case CLOCK_EVT_MODE_ONESHOT: | ||
267 | sh_tmu_disable(p); | ||
268 | disabled = 1; | ||
269 | break; | ||
270 | default: | ||
271 | break; | ||
272 | } | ||
273 | |||
274 | switch (mode) { | ||
275 | case CLOCK_EVT_MODE_PERIODIC: | ||
276 | pr_info("sh_tmu: %s used for periodic clock events\n", | ||
277 | ced->name); | ||
278 | sh_tmu_clock_event_start(p, 1); | ||
279 | break; | ||
280 | case CLOCK_EVT_MODE_ONESHOT: | ||
281 | pr_info("sh_tmu: %s used for oneshot clock events\n", | ||
282 | ced->name); | ||
283 | sh_tmu_clock_event_start(p, 0); | ||
284 | break; | ||
285 | case CLOCK_EVT_MODE_UNUSED: | ||
286 | if (!disabled) | ||
287 | sh_tmu_disable(p); | ||
288 | break; | ||
289 | case CLOCK_EVT_MODE_SHUTDOWN: | ||
290 | default: | ||
291 | break; | ||
292 | } | ||
293 | } | ||
294 | |||
295 | static int sh_tmu_clock_event_next(unsigned long delta, | ||
296 | struct clock_event_device *ced) | ||
297 | { | ||
298 | struct sh_tmu_priv *p = ced_to_sh_tmu(ced); | ||
299 | |||
300 | BUG_ON(ced->mode != CLOCK_EVT_MODE_ONESHOT); | ||
301 | |||
302 | /* program new delta value */ | ||
303 | sh_tmu_set_next(p, delta, 0); | ||
304 | return 0; | ||
305 | } | ||
306 | |||
307 | static void sh_tmu_register_clockevent(struct sh_tmu_priv *p, | ||
308 | char *name, unsigned long rating) | ||
309 | { | ||
310 | struct clock_event_device *ced = &p->ced; | ||
311 | int ret; | ||
312 | |||
313 | memset(ced, 0, sizeof(*ced)); | ||
314 | |||
315 | ced->name = name; | ||
316 | ced->features = CLOCK_EVT_FEAT_PERIODIC; | ||
317 | ced->features |= CLOCK_EVT_FEAT_ONESHOT; | ||
318 | ced->rating = rating; | ||
319 | ced->cpumask = cpumask_of(0); | ||
320 | ced->set_next_event = sh_tmu_clock_event_next; | ||
321 | ced->set_mode = sh_tmu_clock_event_mode; | ||
322 | |||
323 | ret = setup_irq(p->irqaction.irq, &p->irqaction); | ||
324 | if (ret) { | ||
325 | pr_err("sh_tmu: failed to request irq %d\n", | ||
326 | p->irqaction.irq); | ||
327 | return; | ||
328 | } | ||
329 | |||
330 | pr_info("sh_tmu: %s used for clock events\n", ced->name); | ||
331 | clockevents_register_device(ced); | ||
332 | } | ||
333 | |||
334 | static int sh_tmu_register(struct sh_tmu_priv *p, char *name, | ||
335 | unsigned long clockevent_rating, | ||
336 | unsigned long clocksource_rating) | ||
337 | { | ||
338 | if (clockevent_rating) | ||
339 | sh_tmu_register_clockevent(p, name, clockevent_rating); | ||
340 | else if (clocksource_rating) | ||
341 | sh_tmu_register_clocksource(p, name, clocksource_rating); | ||
342 | |||
343 | return 0; | ||
344 | } | ||
345 | |||
346 | static int sh_tmu_setup(struct sh_tmu_priv *p, struct platform_device *pdev) | ||
347 | { | ||
348 | struct sh_timer_config *cfg = pdev->dev.platform_data; | ||
349 | struct resource *res; | ||
350 | int irq, ret; | ||
351 | ret = -ENXIO; | ||
352 | |||
353 | memset(p, 0, sizeof(*p)); | ||
354 | p->pdev = pdev; | ||
355 | |||
356 | if (!cfg) { | ||
357 | dev_err(&p->pdev->dev, "missing platform data\n"); | ||
358 | goto err0; | ||
359 | } | ||
360 | |||
361 | platform_set_drvdata(pdev, p); | ||
362 | |||
363 | res = platform_get_resource(p->pdev, IORESOURCE_MEM, 0); | ||
364 | if (!res) { | ||
365 | dev_err(&p->pdev->dev, "failed to get I/O memory\n"); | ||
366 | goto err0; | ||
367 | } | ||
368 | |||
369 | irq = platform_get_irq(p->pdev, 0); | ||
370 | if (irq < 0) { | ||
371 | dev_err(&p->pdev->dev, "failed to get irq\n"); | ||
372 | goto err0; | ||
373 | } | ||
374 | |||
375 | /* map memory, let mapbase point to our channel */ | ||
376 | p->mapbase = ioremap_nocache(res->start, resource_size(res)); | ||
377 | if (p->mapbase == NULL) { | ||
378 | pr_err("sh_tmu: failed to remap I/O memory\n"); | ||
379 | goto err0; | ||
380 | } | ||
381 | |||
382 | /* setup data for setup_irq() (too early for request_irq()) */ | ||
383 | p->irqaction.name = cfg->name; | ||
384 | p->irqaction.handler = sh_tmu_interrupt; | ||
385 | p->irqaction.dev_id = p; | ||
386 | p->irqaction.irq = irq; | ||
387 | p->irqaction.flags = IRQF_DISABLED | IRQF_TIMER | IRQF_IRQPOLL; | ||
388 | p->irqaction.mask = CPU_MASK_NONE; | ||
389 | |||
390 | /* get hold of clock */ | ||
391 | p->clk = clk_get(&p->pdev->dev, cfg->clk); | ||
392 | if (IS_ERR(p->clk)) { | ||
393 | pr_err("sh_tmu: cannot get clock \"%s\"\n", cfg->clk); | ||
394 | ret = PTR_ERR(p->clk); | ||
395 | goto err1; | ||
396 | } | ||
397 | |||
398 | return sh_tmu_register(p, cfg->name, | ||
399 | cfg->clockevent_rating, | ||
400 | cfg->clocksource_rating); | ||
401 | err1: | ||
402 | iounmap(p->mapbase); | ||
403 | err0: | ||
404 | return ret; | ||
405 | } | ||
406 | |||
407 | static int __devinit sh_tmu_probe(struct platform_device *pdev) | ||
408 | { | ||
409 | struct sh_tmu_priv *p = platform_get_drvdata(pdev); | ||
410 | struct sh_timer_config *cfg = pdev->dev.platform_data; | ||
411 | int ret; | ||
412 | |||
413 | if (p) { | ||
414 | pr_info("sh_tmu: %s kept as earlytimer\n", cfg->name); | ||
415 | return 0; | ||
416 | } | ||
417 | |||
418 | p = kmalloc(sizeof(*p), GFP_KERNEL); | ||
419 | if (p == NULL) { | ||
420 | dev_err(&pdev->dev, "failed to allocate driver data\n"); | ||
421 | return -ENOMEM; | ||
422 | } | ||
423 | |||
424 | ret = sh_tmu_setup(p, pdev); | ||
425 | if (ret) { | ||
426 | kfree(p); | ||
427 | platform_set_drvdata(pdev, NULL); | ||
428 | } | ||
429 | return ret; | ||
430 | } | ||
431 | |||
432 | static int __devexit sh_tmu_remove(struct platform_device *pdev) | ||
433 | { | ||
434 | return -EBUSY; /* cannot unregister clockevent and clocksource */ | ||
435 | } | ||
436 | |||
437 | static struct platform_driver sh_tmu_device_driver = { | ||
438 | .probe = sh_tmu_probe, | ||
439 | .remove = __devexit_p(sh_tmu_remove), | ||
440 | .driver = { | ||
441 | .name = "sh_tmu", | ||
442 | } | ||
443 | }; | ||
444 | |||
445 | static int __init sh_tmu_init(void) | ||
446 | { | ||
447 | return platform_driver_register(&sh_tmu_device_driver); | ||
448 | } | ||
449 | |||
450 | static void __exit sh_tmu_exit(void) | ||
451 | { | ||
452 | platform_driver_unregister(&sh_tmu_device_driver); | ||
453 | } | ||
454 | |||
455 | early_platform_init("earlytimer", &sh_tmu_device_driver); | ||
456 | module_init(sh_tmu_init); | ||
457 | module_exit(sh_tmu_exit); | ||
458 | |||
459 | MODULE_AUTHOR("Magnus Damm"); | ||
460 | MODULE_DESCRIPTION("SuperH TMU Timer Driver"); | ||
461 | MODULE_LICENSE("GPL v2"); | ||