/* * Versatile Express Core Tile Cortex A9x4 Support */ #include <linux/init.h> #include <linux/gfp.h> #include <linux/device.h> #include <linux/dma-mapping.h> #include <linux/platform_device.h> #include <linux/amba/bus.h> #include <linux/amba/clcd.h> #include <linux/platform_data/video-clcd-versatile.h> #include <linux/clkdev.h> #include <linux/vexpress.h> #include <linux/irqchip/arm-gic.h> #include <asm/hardware/arm_timer.h> #include <asm/hardware/cache-l2x0.h> #include <asm/smp_scu.h> #include <asm/smp_twd.h> #include <mach/ct-ca9x4.h> #include <asm/hardware/timer-sp.h> #include <asm/mach/map.h> #include <asm/mach/time.h> #include "core.h" #include <mach/motherboard.h> #include <mach/irqs.h> static struct map_desc ct_ca9x4_io_desc[] __initdata = { { .virtual = V2T_PERIPH, .pfn = __phys_to_pfn(CT_CA9X4_MPIC), .length = SZ_8K, .type = MT_DEVICE, }, }; static void __init ct_ca9x4_map_io(void) { iotable_init(ct_ca9x4_io_desc, ARRAY_SIZE(ct_ca9x4_io_desc)); } static void __init ca9x4_l2_init(void) { #ifdef CONFIG_CACHE_L2X0 void __iomem *l2x0_base = ioremap(CT_CA9X4_L2CC, SZ_4K); if (l2x0_base) { /* set RAM latencies to 1 cycle for this core tile. */ writel(0, l2x0_base + L310_TAG_LATENCY_CTRL); writel(0, l2x0_base + L310_DATA_LATENCY_CTRL); l2x0_init(l2x0_base, 0x00400000, 0xfe0fffff); } else { pr_err("L2C: unable to map L2 cache controller\n"); } #endif } #ifdef CONFIG_HAVE_ARM_TWD static DEFINE_TWD_LOCAL_TIMER(twd_local_timer, A9_MPCORE_TWD, IRQ_LOCALTIMER); static void __init ca9x4_twd_init(void) { int err = twd_local_timer_register(&twd_local_timer); if (err) pr_err("twd_local_timer_register failed %d\n", err); } #else #define ca9x4_twd_init() do {} while(0) #endif static void __init ct_ca9x4_init_irq(void) { gic_init(0, 29, ioremap(A9_MPCORE_GIC_DIST, SZ_4K), ioremap(A9_MPCORE_GIC_CPU, SZ_256)); ca9x4_twd_init(); ca9x4_l2_init(); } static int ct_ca9x4_clcd_setup(struct clcd_fb *fb) { unsigned long framesize = 1024 * 768 * 2; fb->panel = versatile_clcd_get_panel("XVGA"); if (!fb->panel) return -EINVAL; return versatile_clcd_setup_dma(fb, framesize); } static struct clcd_board ct_ca9x4_clcd_data = { .name = "CT-CA9X4", .caps = CLCD_CAP_5551 | CLCD_CAP_565, .check = clcdfb_check, .decode = clcdfb_decode, .setup = ct_ca9x4_clcd_setup, .mmap = versatile_clcd_mmap_dma, .remove = versatile_clcd_remove_dma, }; static AMBA_AHB_DEVICE(clcd, "ct:clcd", 0, CT_CA9X4_CLCDC, IRQ_CT_CA9X4_CLCDC, &ct_ca9x4_clcd_data); static AMBA_APB_DEVICE(dmc, "ct:dmc", 0, CT_CA9X4_DMC, IRQ_CT_CA9X4_DMC, NULL); static AMBA_APB_DEVICE(smc, "ct:smc", 0, CT_CA9X4_SMC, IRQ_CT_CA9X4_SMC, NULL); static AMBA_APB_DEVICE(gpio, "ct:gpio", 0, CT_CA9X4_GPIO, IRQ_CT_CA9X4_GPIO, NULL); static struct amba_device *ct_ca9x4_amba_devs[] __initdata = { &clcd_device, &dmc_device, &smc_device, &gpio_device, }; static struct resource pmu_resources[] = { [0] = { .start = IRQ_CT_CA9X4_PMU_CPU0, .end = IRQ_CT_CA9X4_PMU_CPU0, .flags = IORESOURCE_IRQ, }, [1] = { .start = IRQ_CT_CA9X4_PMU_CPU1, .end = IRQ_CT_CA9X4_PMU_CPU1, .flags = IORESOURCE_IRQ, }, [2] = { .start = IRQ_CT_CA9X4_PMU_CPU2, .end = IRQ_CT_CA9X4_PMU_CPU2, .flags = IORESOURCE_IRQ, }, [3] = { .start = IRQ_CT_CA9X4_PMU_CPU3, .end = IRQ_CT_CA9X4_PMU_CPU3, .flags = IORESOURCE_IRQ, }, }; static struct platform_device pmu_device = { .name = "arm-pmu", .id = -1, .num_resources = ARRAY_SIZE(pmu_resources), .resource = pmu_resources, }; static struct clk_lookup osc1_lookup = { .dev_id = "ct:clcd", }; static struct platform_device osc1_device = { .name = "vexpress-osc", .id = 1, .num_resources = 1, .resource = (struct resource []) { VEXPRESS_RES_FUNC(0xf, 1), }, .dev.platform_data = &osc1_lookup, }; static void __init ct_ca9x4_init(void) { int i; for (i = 0; i < ARRAY_SIZE(ct_ca9x4_amba_devs); i++) amba_device_register(ct_ca9x4_amba_devs[i], &iomem_resource); platform_device_register(&pmu_device); vexpress_syscfg_device_register(&osc1_device); } #ifdef CONFIG_SMP static void *ct_ca9x4_scu_base __initdata; static void __init ct_ca9x4_init_cpu_map(void) { int i, ncores; ct_ca9x4_scu_base = ioremap(A9_MPCORE_SCU, SZ_128); if (WARN_ON(!ct_ca9x4_scu_base)) return; ncores = scu_get_core_count(ct_ca9x4_scu_base); if (ncores > nr_cpu_ids) { pr_warn("SMP: %u cores greater than maximum (%u), clipping\n", ncores, nr_cpu_ids); ncores = nr_cpu_ids; } for (i = 0; i < ncores; ++i) set_cpu_possible(i, true); } static void __init ct_ca9x4_smp_enable(unsigned int max_cpus) { scu_enable(ct_ca9x4_scu_base); } #endif struct ct_desc ct_ca9x4_desc __initdata = { .id = V2M_CT_ID_CA9, .name = "CA9x4", .map_io = ct_ca9x4_map_io, .init_irq = ct_ca9x4_init_irq, .init_tile = ct_ca9x4_init, #ifdef CONFIG_SMP .init_cpu_map = ct_ca9x4_init_cpu_map, .smp_enable = ct_ca9x4_smp_enable, #endif };