Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1 | /* SPDX-License-Identifier: GPL-2.0 */ |
| 2 | #ifndef __ASM_SH_PAGE_H |
| 3 | #define __ASM_SH_PAGE_H |
| 4 | |
| 5 | /* |
| 6 | * Copyright (C) 1999 Niibe Yutaka |
| 7 | */ |
| 8 | |
| 9 | #include <linux/const.h> |
| 10 | |
| 11 | /* PAGE_SHIFT determines the page size */ |
| 12 | #if defined(CONFIG_PAGE_SIZE_4KB) |
| 13 | # define PAGE_SHIFT 12 |
| 14 | #elif defined(CONFIG_PAGE_SIZE_8KB) |
| 15 | # define PAGE_SHIFT 13 |
| 16 | #elif defined(CONFIG_PAGE_SIZE_16KB) |
| 17 | # define PAGE_SHIFT 14 |
| 18 | #elif defined(CONFIG_PAGE_SIZE_64KB) |
| 19 | # define PAGE_SHIFT 16 |
| 20 | #else |
| 21 | # error "Bogus kernel page size?" |
| 22 | #endif |
| 23 | |
| 24 | #define PAGE_SIZE (_AC(1, UL) << PAGE_SHIFT) |
| 25 | #define PAGE_MASK (~(PAGE_SIZE-1)) |
| 26 | #define PTE_MASK PAGE_MASK |
| 27 | |
| 28 | #if defined(CONFIG_HUGETLB_PAGE_SIZE_64K) |
| 29 | #define HPAGE_SHIFT 16 |
| 30 | #elif defined(CONFIG_HUGETLB_PAGE_SIZE_256K) |
| 31 | #define HPAGE_SHIFT 18 |
| 32 | #elif defined(CONFIG_HUGETLB_PAGE_SIZE_1MB) |
| 33 | #define HPAGE_SHIFT 20 |
| 34 | #elif defined(CONFIG_HUGETLB_PAGE_SIZE_4MB) |
| 35 | #define HPAGE_SHIFT 22 |
| 36 | #elif defined(CONFIG_HUGETLB_PAGE_SIZE_64MB) |
| 37 | #define HPAGE_SHIFT 26 |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 38 | #endif |
| 39 | |
| 40 | #ifdef CONFIG_HUGETLB_PAGE |
| 41 | #define HPAGE_SIZE (1UL << HPAGE_SHIFT) |
| 42 | #define HPAGE_MASK (~(HPAGE_SIZE-1)) |
| 43 | #define HUGETLB_PAGE_ORDER (HPAGE_SHIFT-PAGE_SHIFT) |
| 44 | #endif |
| 45 | |
| 46 | #ifndef __ASSEMBLY__ |
| 47 | #include <asm/uncached.h> |
| 48 | |
| 49 | extern unsigned long shm_align_mask; |
| 50 | extern unsigned long max_low_pfn, min_low_pfn; |
| 51 | extern unsigned long memory_start, memory_end, memory_limit; |
| 52 | |
| 53 | static inline unsigned long |
| 54 | pages_do_alias(unsigned long addr1, unsigned long addr2) |
| 55 | { |
| 56 | return (addr1 ^ addr2) & shm_align_mask; |
| 57 | } |
| 58 | |
| 59 | #define clear_page(page) memset((void *)(page), 0, PAGE_SIZE) |
| 60 | extern void copy_page(void *to, void *from); |
| 61 | #define copy_user_page(to, from, vaddr, pg) __copy_user(to, from, PAGE_SIZE) |
| 62 | |
| 63 | struct page; |
| 64 | struct vm_area_struct; |
| 65 | |
| 66 | extern void copy_user_highpage(struct page *to, struct page *from, |
| 67 | unsigned long vaddr, struct vm_area_struct *vma); |
| 68 | #define __HAVE_ARCH_COPY_USER_HIGHPAGE |
| 69 | extern void clear_user_highpage(struct page *page, unsigned long vaddr); |
| 70 | #define clear_user_highpage clear_user_highpage |
| 71 | |
| 72 | /* |
| 73 | * These are used to make use of C type-checking.. |
| 74 | */ |
| 75 | #ifdef CONFIG_X2TLB |
| 76 | typedef struct { unsigned long pte_low, pte_high; } pte_t; |
| 77 | typedef struct { unsigned long long pgprot; } pgprot_t; |
| 78 | typedef struct { unsigned long long pgd; } pgd_t; |
| 79 | #define pte_val(x) \ |
| 80 | ((x).pte_low | ((unsigned long long)(x).pte_high << 32)) |
| 81 | #define __pte(x) \ |
| 82 | ({ pte_t __pte = {(x), ((unsigned long long)(x)) >> 32}; __pte; }) |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 83 | #else |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 84 | typedef struct { unsigned long pte_low; } pte_t; |
| 85 | typedef struct { unsigned long pgprot; } pgprot_t; |
| 86 | typedef struct { unsigned long pgd; } pgd_t; |
| 87 | #define pte_val(x) ((x).pte_low) |
| 88 | #define __pte(x) ((pte_t) { (x) } ) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 89 | #endif |
| 90 | |
| 91 | #define pgd_val(x) ((x).pgd) |
| 92 | #define pgprot_val(x) ((x).pgprot) |
| 93 | |
| 94 | #define __pgd(x) ((pgd_t) { (x) } ) |
| 95 | #define __pgprot(x) ((pgprot_t) { (x) } ) |
| 96 | |
| 97 | typedef struct page *pgtable_t; |
| 98 | |
| 99 | #define pte_pgprot(x) __pgprot(pte_val(x) & PTE_FLAGS_MASK) |
| 100 | |
| 101 | #endif /* !__ASSEMBLY__ */ |
| 102 | |
| 103 | /* |
| 104 | * __MEMORY_START and SIZE are the physical addresses and size of RAM. |
| 105 | */ |
| 106 | #define __MEMORY_START CONFIG_MEMORY_START |
| 107 | #define __MEMORY_SIZE CONFIG_MEMORY_SIZE |
| 108 | |
| 109 | /* |
| 110 | * PHYSICAL_OFFSET is the offset in physical memory where the base |
| 111 | * of the kernel is loaded. |
| 112 | */ |
| 113 | #ifdef CONFIG_PHYSICAL_START |
| 114 | #define PHYSICAL_OFFSET (CONFIG_PHYSICAL_START - __MEMORY_START) |
| 115 | #else |
| 116 | #define PHYSICAL_OFFSET 0 |
| 117 | #endif |
| 118 | |
| 119 | /* |
| 120 | * PAGE_OFFSET is the virtual address of the start of kernel address |
| 121 | * space. |
| 122 | */ |
| 123 | #define PAGE_OFFSET CONFIG_PAGE_OFFSET |
| 124 | |
| 125 | /* |
| 126 | * Virtual to physical RAM address translation. |
| 127 | * |
| 128 | * In 29 bit mode, the physical offset of RAM from address 0 is visible in |
| 129 | * the kernel virtual address space, and thus we don't have to take |
| 130 | * this into account when translating. However in 32 bit mode this offset |
| 131 | * is not visible (it is part of the PMB mapping) and so needs to be |
| 132 | * added or subtracted as required. |
| 133 | */ |
| 134 | #ifdef CONFIG_PMB |
| 135 | #define ___pa(x) ((x)-PAGE_OFFSET+__MEMORY_START) |
| 136 | #define ___va(x) ((x)+PAGE_OFFSET-__MEMORY_START) |
| 137 | #else |
| 138 | #define ___pa(x) ((x)-PAGE_OFFSET) |
| 139 | #define ___va(x) ((x)+PAGE_OFFSET) |
| 140 | #endif |
| 141 | |
| 142 | #ifndef __ASSEMBLY__ |
| 143 | #define __pa(x) ___pa((unsigned long)x) |
| 144 | #define __va(x) (void *)___va((unsigned long)x) |
| 145 | #endif /* !__ASSEMBLY__ */ |
| 146 | |
| 147 | #ifdef CONFIG_UNCACHED_MAPPING |
| 148 | #if defined(CONFIG_29BIT) |
| 149 | #define UNCAC_ADDR(addr) P2SEGADDR(addr) |
| 150 | #define CAC_ADDR(addr) P1SEGADDR(addr) |
| 151 | #else |
| 152 | #define UNCAC_ADDR(addr) ((addr) - PAGE_OFFSET + uncached_start) |
| 153 | #define CAC_ADDR(addr) ((addr) - uncached_start + PAGE_OFFSET) |
| 154 | #endif |
| 155 | #else |
| 156 | #define UNCAC_ADDR(addr) ((addr)) |
| 157 | #define CAC_ADDR(addr) ((addr)) |
| 158 | #endif |
| 159 | |
| 160 | #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) |
| 161 | #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT) |
| 162 | |
| 163 | /* |
| 164 | * PFN = physical frame number (ie PFN 0 == physical address 0) |
| 165 | * PFN_START is the PFN of the first page of RAM. By defining this we |
| 166 | * don't have struct page entries for the portion of address space |
| 167 | * between physical address 0 and the start of RAM. |
| 168 | */ |
| 169 | #define PFN_START (__MEMORY_START >> PAGE_SHIFT) |
| 170 | #define ARCH_PFN_OFFSET (PFN_START) |
| 171 | #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) |
| 172 | #ifdef CONFIG_FLATMEM |
| 173 | #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_low_pfn) |
| 174 | #endif |
| 175 | #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) |
| 176 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 177 | #include <asm-generic/memory_model.h> |
| 178 | #include <asm-generic/getorder.h> |
| 179 | |
| 180 | /* |
| 181 | * Some drivers need to perform DMA into kmalloc'ed buffers |
| 182 | * and so we have to increase the kmalloc minalign for this. |
| 183 | */ |
| 184 | #define ARCH_DMA_MINALIGN L1_CACHE_BYTES |
| 185 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 186 | #endif /* __ASM_SH_PAGE_H */ |