/* * Blackfin CPLB initialization * * Copyright 2007-2009 Analog Devices Inc. * * Licensed under the GPL-2 or later. */ #include <linux/module.h> #include <asm/blackfin.h> #include <asm/cacheflush.h> #include <asm/cplb.h> #include <asm/cplbinit.h> #include <asm/mem_map.h> struct cplb_entry icplb_tbl[NR_CPUS][MAX_CPLBS] PDT_ATTR; struct cplb_entry dcplb_tbl[NR_CPUS][MAX_CPLBS] PDT_ATTR; int first_switched_icplb PDT_ATTR; int first_switched_dcplb PDT_ATTR; struct cplb_boundary dcplb_bounds[9] PDT_ATTR; struct cplb_boundary icplb_bounds[9] PDT_ATTR; int icplb_nr_bounds PDT_ATTR; int dcplb_nr_bounds PDT_ATTR; void __init generate_cplb_tables_cpu(unsigned int cpu) { int i_d, i_i; unsigned long addr; unsigned long cplb_pageflags, cplb_pagesize; struct cplb_entry *d_tbl = dcplb_tbl[cpu]; struct cplb_entry *i_tbl = icplb_tbl[cpu]; printk(KERN_INFO "NOMPU: setting up cplb tables\n"); i_d = i_i = 0; #ifdef CONFIG_DEBUG_HUNT_FOR_ZERO /* Set up the zero page. */ d_tbl[i_d].addr = 0; d_tbl[i_d++].data = SDRAM_OOPS | PAGE_SIZE_1KB; i_tbl[i_i].addr = 0; i_tbl[i_i++].data = SDRAM_OOPS | PAGE_SIZE_1KB; #endif /* Cover kernel memory with 4M pages. */ addr = 0; #ifdef PAGE_SIZE_16MB cplb_pageflags = PAGE_SIZE_16MB; cplb_pagesize = SIZE_16M; #else cplb_pageflags = PAGE_SIZE_4MB; cplb_pagesize = SIZE_4M; #endif for (; addr < memory_start; addr += cplb_pagesize) { d_tbl[i_d].addr = addr; d_tbl[i_d++].data = SDRAM_DGENERIC | cplb_pageflags; i_tbl[i_i].addr = addr; i_tbl[i_i++].data = SDRAM_IGENERIC | cplb_pageflags; } #ifdef CONFIG_ROMKERNEL /* Cover kernel XIP flash area */ #ifdef CONFIG_BF60x addr = CONFIG_ROM_BASE & ~(16 * 1024 * 1024 - 1); d_tbl[i_d].addr = addr; d_tbl[i_d++].data = SDRAM_DGENERIC | PAGE_SIZE_16MB; i_tbl[i_i].addr = addr; i_tbl[i_i++].data = SDRAM_IGENERIC | PAGE_SIZE_16MB; #else addr = CONFIG_ROM_BASE & ~(4 * 1024 * 1024 - 1); d_tbl[i_d].addr = addr; d_tbl[i_d++].data = SDRAM_DGENERIC | PAGE_SIZE_4MB; i_tbl[i_i].addr = addr; i_tbl[i_i++].data = SDRAM_IGENERIC | PAGE_SIZE_4MB; #endif #endif /* Cover L1 memory. One 4M area for code and data each is enough. */ if (cpu == 0) { if (L1_DATA_A_LENGTH || L1_DATA_B_LENGTH) { d_tbl[i_d].addr = L1_DATA_A_START; d_tbl[i_d++].data = L1_DMEMORY | PAGE_SIZE_4MB; } i_tbl[i_i].addr = L1_CODE_START; i_tbl[i_i++].data = L1_IMEMORY | PAGE_SIZE_4MB; } #ifdef CONFIG_SMP else { if (L1_DATA_A_LENGTH || L1_DATA_B_LENGTH) { d_tbl[i_d].addr = COREB_L1_DATA_A_START; d_tbl[i_d++].data = L1_DMEMORY | PAGE_SIZE_4MB; } i_tbl[i_i].addr = COREB_L1_CODE_START; i_tbl[i_i++].data = L1_IMEMORY | PAGE_SIZE_4MB; } #endif first_switched_dcplb = i_d; first_switched_icplb = i_i; BUG_ON(first_switched_dcplb > MAX_CPLBS); BUG_ON(first_switched_icplb > MAX_CPLBS); while (i_d < MAX_CPLBS) d_tbl[i_d++].data = 0; while (i_i < MAX_CPLBS) i_tbl[i_i++].data = 0; } void __init generate_cplb_tables_all(void) { unsigned long uncached_end; int i_d, i_i; i_d = 0; /* Normal RAM, including MTD FS. */ #ifdef CONFIG_MTD_UCLINUX uncached_end = memory_mtd_start + mtd_size; #else uncached_end = memory_end; #endif /* * if DMA uncached is less than 1MB, mark the 1MB chunk as uncached * so that we don't have to use 4kB pages and cause CPLB thrashing */ if ((DMA_UNCACHED_REGION >= 1 * 1024 * 1024) || !DMA_UNCACHED_REGION || ((_ramend - uncached_end) >= 1 * 1024 * 1024)) dcplb_bounds[i_d].eaddr = uncached_end; else dcplb_bounds[i_d].eaddr = uncached_end & ~(1 * 1024 * 1024 - 1); dcplb_bounds[i_d++].data = SDRAM_DGENERIC; /* DMA uncached region. */ if (DMA_UNCACHED_REGION) { dcplb_bounds[i_d].eaddr = _ramend; dcplb_bounds[i_d++].data = SDRAM_DNON_CHBL; } if (_ramend != physical_mem_end) { /* Reserved memory. */ dcplb_bounds[i_d].eaddr = physical_mem_end; dcplb_bounds[i_d++].data = (reserved_mem_dcache_on ? SDRAM_DGENERIC : SDRAM_DNON_CHBL); } /* Addressing hole up to the async bank. */ dcplb_bounds[i_d].eaddr = ASYNC_BANK0_BASE; dcplb_bounds[i_d++].data = 0; /* ASYNC banks. */ dcplb_bounds[i_d].eaddr = ASYNC_BANK3_BASE + ASYNC_BANK3_SIZE; dcplb_bounds[i_d++].data = SDRAM_EBIU; /* Addressing hole up to BootROM. */ dcplb_bounds[i_d].eaddr = BOOT_ROM_START; dcplb_bounds[i_d++].data = 0; /* BootROM -- largest one should be less than 1 meg. */ dcplb_bounds[i_d].eaddr = BOOT_ROM_START + BOOT_ROM_LENGTH; dcplb_bounds[i_d++].data = SDRAM_DGENERIC; if (L2_LENGTH) { /* Addressing hole up to L2 SRAM. */ dcplb_bounds[i_d].eaddr = L2_START; dcplb_bounds[i_d++].data = 0; /* L2 SRAM. */ dcplb_bounds[i_d].eaddr = L2_START + L2_LENGTH; dcplb_bounds[i_d++].data = L2_DMEMORY; } dcplb_nr_bounds = i_d; BUG_ON(dcplb_nr_bounds > ARRAY_SIZE(dcplb_bounds)); i_i = 0; /* Normal RAM, including MTD FS. */ icplb_bounds[i_i].eaddr = uncached_end; icplb_bounds[i_i++].data = SDRAM_IGENERIC; if (_ramend != physical_mem_end) { /* DMA uncached region. */ if (DMA_UNCACHED_REGION) { /* Normally this hole is caught by the async below. */ icplb_bounds[i_i].eaddr = _ramend; icplb_bounds[i_i++].data = 0; } /* Reserved memory. */ icplb_bounds[i_i].eaddr = physical_mem_end; icplb_bounds[i_i++].data = (reserved_mem_icache_on ? SDRAM_IGENERIC : SDRAM_INON_CHBL); } /* Addressing hole up to the async bank. */ icplb_bounds[i_i].eaddr = ASYNC_BANK0_BASE; icplb_bounds[i_i++].data = 0; /* ASYNC banks. */ icplb_bounds[i_i].eaddr = ASYNC_BANK3_BASE + ASYNC_BANK3_SIZE; icplb_bounds[i_i++].data = SDRAM_EBIU; /* Addressing hole up to BootROM. */ icplb_bounds[i_i].eaddr = BOOT_ROM_START; icplb_bounds[i_i++].data = 0; /* BootROM -- largest one should be less than 1 meg. */ icplb_bounds[i_i].eaddr = BOOT_ROM_START + BOOT_ROM_LENGTH; icplb_bounds[i_i++].data = SDRAM_IGENERIC; if (L2_LENGTH) { /* Addressing hole up to L2 SRAM. */ icplb_bounds[i_i].eaddr = L2_START; icplb_bounds[i_i++].data = 0; /* L2 SRAM. */ icplb_bounds[i_i].eaddr = L2_START + L2_LENGTH; icplb_bounds[i_i++].data = L2_IMEMORY; } icplb_nr_bounds = i_i; BUG_ON(icplb_nr_bounds > ARRAY_SIZE(icplb_bounds)); }