1 | #ifndef _ASM_X86_PGTABLE_32_AREAS_H |
2 | #define _ASM_X86_PGTABLE_32_AREAS_H |
3 | |
4 | #include <asm/cpu_entry_area.h> |
5 | |
6 | /* |
7 | * Just any arbitrary offset to the start of the vmalloc VM area: the |
8 | * current 8MB value just means that there will be a 8MB "hole" after the |
9 | * physical memory until the kernel virtual memory starts. That means that |
10 | * any out-of-bounds memory accesses will hopefully be caught. |
11 | * The vmalloc() routines leaves a hole of 4kB between each vmalloced |
12 | * area for the same reason. ;) |
13 | */ |
14 | #define VMALLOC_OFFSET (8 * 1024 * 1024) |
15 | |
16 | #ifndef __ASSEMBLY__ |
17 | extern bool __vmalloc_start_set; /* set once high_memory is set */ |
18 | #endif |
19 | |
20 | #define VMALLOC_START ((unsigned long)high_memory + VMALLOC_OFFSET) |
21 | #ifdef CONFIG_X86_PAE |
22 | #define LAST_PKMAP 512 |
23 | #else |
24 | #define LAST_PKMAP 1024 |
25 | #endif |
26 | |
27 | #define CPU_ENTRY_AREA_PAGES (NR_CPUS * DIV_ROUND_UP(sizeof(struct cpu_entry_area), PAGE_SIZE)) |
28 | |
29 | /* The +1 is for the readonly IDT page: */ |
30 | #define CPU_ENTRY_AREA_BASE \ |
31 | ((FIXADDR_TOT_START - PAGE_SIZE*(CPU_ENTRY_AREA_PAGES+1)) & PMD_MASK) |
32 | |
33 | #define LDT_BASE_ADDR \ |
34 | ((CPU_ENTRY_AREA_BASE - PAGE_SIZE) & PMD_MASK) |
35 | |
36 | #define LDT_END_ADDR (LDT_BASE_ADDR + PMD_SIZE) |
37 | |
38 | #define PKMAP_BASE \ |
39 | ((LDT_BASE_ADDR - PAGE_SIZE) & PMD_MASK) |
40 | |
41 | #ifdef CONFIG_HIGHMEM |
42 | # define VMALLOC_END (PKMAP_BASE - 2 * PAGE_SIZE) |
43 | #else |
44 | # define VMALLOC_END (LDT_BASE_ADDR - 2 * PAGE_SIZE) |
45 | #endif |
46 | |
47 | #define MODULES_VADDR VMALLOC_START |
48 | #define MODULES_END VMALLOC_END |
49 | #define MODULES_LEN (MODULES_VADDR - MODULES_END) |
50 | |
51 | #define MAXMEM (VMALLOC_END - PAGE_OFFSET - __VMALLOC_RESERVE) |
52 | |
53 | #endif /* _ASM_X86_PGTABLE_32_AREAS_H */ |
54 | |