|
37 | 37 |
|
38 | 38 | #ifndef __ASSEMBLY__ |
39 | 39 |
|
40 | | -#define PAGE_UP(addr) (((addr)+((PAGE_SIZE)-1))&(~((PAGE_SIZE)-1))) |
41 | | -#define PAGE_DOWN(addr) ((addr)&(~((PAGE_SIZE)-1))) |
42 | | - |
43 | | -/* align addr on a size boundary - adjust address up/down if needed */ |
44 | | -#define _ALIGN_UP(addr, size) (((addr)+((size)-1))&(~((size)-1))) |
45 | | -#define _ALIGN_DOWN(addr, size) ((addr)&(~((size)-1))) |
46 | | - |
47 | | -/* align addr on a size boundary - adjust address up if needed */ |
48 | | -#define _ALIGN(addr, size) _ALIGN_UP(addr, size) |
49 | | - |
50 | 40 | #define clear_page(pgaddr) memset((pgaddr), 0, PAGE_SIZE) |
51 | 41 | #define copy_page(to, from) memcpy((to), (from), PAGE_SIZE) |
52 | 42 |
|
@@ -89,59 +79,68 @@ typedef struct page *pgtable_t; |
89 | 79 | #endif |
90 | 80 |
|
91 | 81 | #ifdef CONFIG_MMU |
92 | | -extern unsigned long va_pa_offset; |
93 | | -#ifdef CONFIG_64BIT |
94 | | -extern unsigned long va_kernel_pa_offset; |
95 | | -#endif |
96 | | -#ifdef CONFIG_XIP_KERNEL |
97 | | -extern unsigned long va_kernel_xip_pa_offset; |
98 | | -#endif |
99 | 82 | extern unsigned long pfn_base; |
100 | 83 | #define ARCH_PFN_OFFSET (pfn_base) |
101 | 84 | #else |
102 | | -#define va_pa_offset 0 |
103 | | -#ifdef CONFIG_64BIT |
104 | | -#define va_kernel_pa_offset 0 |
105 | | -#endif |
106 | 85 | #define ARCH_PFN_OFFSET (PAGE_OFFSET >> PAGE_SHIFT) |
107 | 86 | #endif /* CONFIG_MMU */ |
108 | 87 |
|
109 | | -extern unsigned long kernel_virt_addr; |
110 | | - |
| 88 | +struct kernel_mapping { |
| 89 | + unsigned long virt_addr; |
| 90 | + uintptr_t phys_addr; |
| 91 | + uintptr_t size; |
| 92 | + /* Offset between linear mapping virtual address and kernel load address */ |
| 93 | + unsigned long va_pa_offset; |
111 | 94 | #ifdef CONFIG_64BIT |
112 | | -#define linear_mapping_pa_to_va(x) ((void *)((unsigned long)(x) + va_pa_offset)) |
| 95 | + /* Offset between kernel mapping virtual address and kernel load address */ |
| 96 | + unsigned long va_kernel_pa_offset; |
| 97 | +#endif |
| 98 | + unsigned long va_kernel_xip_pa_offset; |
113 | 99 | #ifdef CONFIG_XIP_KERNEL |
| 100 | + uintptr_t xiprom; |
| 101 | + uintptr_t xiprom_sz; |
| 102 | +#endif |
| 103 | +}; |
| 104 | + |
| 105 | +extern struct kernel_mapping kernel_map; |
| 106 | + |
| 107 | +#ifdef CONFIG_64BIT |
| 108 | +#define is_kernel_mapping(x) \ |
| 109 | + ((x) >= kernel_map.virt_addr && (x) < (kernel_map.virt_addr + kernel_map.size)) |
| 110 | +#define is_linear_mapping(x) \ |
| 111 | + ((x) >= PAGE_OFFSET && (x) < kernel_map.virt_addr) |
| 112 | + |
| 113 | +#define linear_mapping_pa_to_va(x) ((void *)((unsigned long)(x) + kernel_map.va_pa_offset)) |
114 | 114 | #define kernel_mapping_pa_to_va(y) ({ \ |
115 | 115 | unsigned long _y = y; \ |
116 | 116 | (_y >= CONFIG_PHYS_RAM_BASE) ? \ |
117 | | - (void *)((unsigned long)(_y) + va_kernel_pa_offset + XIP_OFFSET) : \ |
118 | | - (void *)((unsigned long)(_y) + va_kernel_xip_pa_offset); \ |
| 117 | + (void *)((unsigned long)(_y) + kernel_map.va_kernel_pa_offset + XIP_OFFSET) : \ |
| 118 | + (void *)((unsigned long)(_y) + kernel_map.va_kernel_xip_pa_offset); \ |
119 | 119 | }) |
120 | | -#else |
121 | | -#define kernel_mapping_pa_to_va(x) ((void *)((unsigned long)(x) + va_kernel_pa_offset)) |
122 | | -#endif |
123 | 120 | #define __pa_to_va_nodebug(x) linear_mapping_pa_to_va(x) |
124 | 121 |
|
125 | | -#define linear_mapping_va_to_pa(x) ((unsigned long)(x) - va_pa_offset) |
126 | | -#ifdef CONFIG_XIP_KERNEL |
| 122 | +#define linear_mapping_va_to_pa(x) ((unsigned long)(x) - kernel_map.va_pa_offset) |
127 | 123 | #define kernel_mapping_va_to_pa(y) ({ \ |
128 | 124 | unsigned long _y = y; \ |
129 | | - (_y < kernel_virt_addr + XIP_OFFSET) ? \ |
130 | | - ((unsigned long)(_y) - va_kernel_xip_pa_offset) : \ |
131 | | - ((unsigned long)(_y) - va_kernel_pa_offset - XIP_OFFSET); \ |
| 125 | + (_y < kernel_map.virt_addr + XIP_OFFSET) ? \ |
| 126 | + ((unsigned long)(_y) - kernel_map.va_kernel_xip_pa_offset) : \ |
| 127 | + ((unsigned long)(_y) - kernel_map.va_kernel_pa_offset - XIP_OFFSET); \ |
132 | 128 | }) |
133 | | -#else |
134 | | -#define kernel_mapping_va_to_pa(x) ((unsigned long)(x) - va_kernel_pa_offset) |
135 | | -#endif |
| 129 | + |
136 | 130 | #define __va_to_pa_nodebug(x) ({ \ |
137 | 131 | unsigned long _x = x; \ |
138 | | - (_x < kernel_virt_addr) ? \ |
| 132 | + is_linear_mapping(_x) ? \ |
139 | 133 | linear_mapping_va_to_pa(_x) : kernel_mapping_va_to_pa(_x); \ |
140 | 134 | }) |
141 | 135 | #else |
142 | | -#define __pa_to_va_nodebug(x) ((void *)((unsigned long) (x) + va_pa_offset)) |
143 | | -#define __va_to_pa_nodebug(x) ((unsigned long)(x) - va_pa_offset) |
144 | | -#endif |
| 136 | +#define is_kernel_mapping(x) \ |
| 137 | + ((x) >= kernel_map.virt_addr && (x) < (kernel_map.virt_addr + kernel_map.size)) |
| 138 | +#define is_linear_mapping(x) \ |
| 139 | + ((x) >= PAGE_OFFSET) |
| 140 | + |
| 141 | +#define __pa_to_va_nodebug(x) ((void *)((unsigned long) (x) + kernel_map.va_pa_offset)) |
| 142 | +#define __va_to_pa_nodebug(x) ((unsigned long)(x) - kernel_map.va_pa_offset) |
| 143 | +#endif /* CONFIG_64BIT */ |
145 | 144 |
|
146 | 145 | #ifdef CONFIG_DEBUG_VIRTUAL |
147 | 146 | extern phys_addr_t __virt_to_phys(unsigned long x); |
|
0 commit comments