^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) /* SPDX-License-Identifier: GPL-2.0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) #ifndef _ASM_IA64_PAGE_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) #define _ASM_IA64_PAGE_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) * Pagetable related stuff.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) * Copyright (C) 1998, 1999, 2002 Hewlett-Packard Co
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) * David Mosberger-Tang <davidm@hpl.hp.com>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) #include <asm/intrinsics.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) #include <asm/types.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) * The top three bits of an IA64 address are its Region Number.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) * Different regions are assigned to different purposes.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) #define RGN_SHIFT (61)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) #define RGN_BASE(r) (__IA64_UL_CONST(r)<<RGN_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20) #define RGN_BITS (RGN_BASE(-1))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22) #define RGN_KERNEL 7 /* Identity mapped region */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23) #define RGN_UNCACHED 6 /* Identity mapped I/O region */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) #define RGN_GATE 5 /* Gate page, Kernel text, etc */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) #define RGN_HPAGE 4 /* For Huge TLB pages */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) * PAGE_SHIFT determines the actual kernel page size.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30) #if defined(CONFIG_IA64_PAGE_SIZE_4KB)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) # define PAGE_SHIFT 12
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) #elif defined(CONFIG_IA64_PAGE_SIZE_8KB)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) # define PAGE_SHIFT 13
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34) #elif defined(CONFIG_IA64_PAGE_SIZE_16KB)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35) # define PAGE_SHIFT 14
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) #elif defined(CONFIG_IA64_PAGE_SIZE_64KB)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) # define PAGE_SHIFT 16
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39) # error Unsupported page size!
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) #define PAGE_SIZE (__IA64_UL_CONST(1) << PAGE_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) #define PAGE_MASK (~(PAGE_SIZE - 1))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45) #define PERCPU_PAGE_SHIFT 18 /* log2() of max. size of per-CPU area */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) #define PERCPU_PAGE_SIZE (__IA64_UL_CONST(1) << PERCPU_PAGE_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) #ifdef CONFIG_HUGETLB_PAGE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50) # define HPAGE_REGION_BASE RGN_BASE(RGN_HPAGE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) # define HPAGE_SHIFT hpage_shift
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) # define HPAGE_SHIFT_DEFAULT 28 /* check ia64 SDM for architecture supported size */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53) # define HPAGE_SIZE (__IA64_UL_CONST(1) << HPAGE_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54) # define HPAGE_MASK (~(HPAGE_SIZE - 1))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) # define HAVE_ARCH_HUGETLB_UNMAPPED_AREA
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) #endif /* CONFIG_HUGETLB_PAGE */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) #ifdef __ASSEMBLY__
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60) # define __pa(x) ((x) - PAGE_OFFSET)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61) # define __va(x) ((x) + PAGE_OFFSET)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) #else /* !__ASSEMBLY */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63) # define STRICT_MM_TYPECHECKS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) extern void clear_page (void *page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66) extern void copy_page (void *to, void *from);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) * clear_user_page() and copy_user_page() can't be inline functions because
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70) * flush_dcache_page() can't be defined until later...
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) #define clear_user_page(addr, vaddr, page) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73) do { \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) clear_page(addr); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75) flush_dcache_page(page); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76) } while (0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78) #define copy_user_page(to, from, vaddr, page) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79) do { \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) copy_page((to), (from)); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81) flush_dcache_page(page); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82) } while (0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85) #define alloc_zeroed_user_highpage_movable(vma, vaddr) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86) ({ \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87) struct page *page = alloc_page_vma( \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) GFP_HIGHUSER_MOVABLE | __GFP_ZERO, vma, vaddr); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) if (page) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) flush_dcache_page(page); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91) page; \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) })
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94) #define __HAVE_ARCH_ALLOC_ZEROED_USER_HIGHPAGE_MOVABLE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96) #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) #ifdef CONFIG_VIRTUAL_MEM_MAP
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99) extern int ia64_pfn_valid (unsigned long pfn);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) # define ia64_pfn_valid(pfn) 1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104) #ifdef CONFIG_VIRTUAL_MEM_MAP
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) extern struct page *vmem_map;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106) #ifdef CONFIG_DISCONTIGMEM
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) # define page_to_pfn(page) ((unsigned long) (page - vmem_map))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) # define pfn_to_page(pfn) (vmem_map + (pfn))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) # define __pfn_to_phys(pfn) PFN_PHYS(pfn)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) # include <asm-generic/memory_model.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) # include <asm-generic/memory_model.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) #ifdef CONFIG_FLATMEM
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) # define pfn_valid(pfn) (((pfn) < max_mapnr) && ia64_pfn_valid(pfn))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) #elif defined(CONFIG_DISCONTIGMEM)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120) extern unsigned long min_low_pfn;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) extern unsigned long max_low_pfn;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) # define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn) && ia64_pfn_valid(pfn))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127) #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) typedef union ia64_va {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) struct {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) unsigned long off : 61; /* intra-region offset */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) unsigned long reg : 3; /* region number */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) } f;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134) unsigned long l;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) void *p;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) } ia64_va;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) * Note: These macros depend on the fact that PAGE_OFFSET has all
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) * region bits set to 1 and all other bits set to zero. They are
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) * expressed in this way to ensure they result in a single "dep"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) * instruction.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) #define __pa(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg = 0; _v.l;})
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) #define __va(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg = -1; _v.p;})
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) #define REGION_NUMBER(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg;})
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) #define REGION_OFFSET(x) ({ia64_va _v; _v.l = (long) (x); _v.f.off;})
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) #ifdef CONFIG_HUGETLB_PAGE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) # define htlbpage_to_page(x) (((unsigned long) REGION_NUMBER(x) << 61) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) | (REGION_OFFSET(x) >> (HPAGE_SHIFT-PAGE_SHIFT)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) # define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) extern unsigned int hpage_shift;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) static __inline__ int
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) get_order (unsigned long size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) long double d = size - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) long order;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) order = ia64_getf_exp(d);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164) order = order - PAGE_SHIFT - 0xffff + 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) if (order < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) order = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) return order;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170) #endif /* !__ASSEMBLY__ */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) #ifdef STRICT_MM_TYPECHECKS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) * These are used to make use of C type-checking..
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) typedef struct { unsigned long pte; } pte_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) typedef struct { unsigned long pmd; } pmd_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) #if CONFIG_PGTABLE_LEVELS == 4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) typedef struct { unsigned long pud; } pud_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) typedef struct { unsigned long pgd; } pgd_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) typedef struct { unsigned long pgprot; } pgprot_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) typedef struct page *pgtable_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) # define pte_val(x) ((x).pte)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) # define pmd_val(x) ((x).pmd)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) #if CONFIG_PGTABLE_LEVELS == 4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) # define pud_val(x) ((x).pud)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) # define pgd_val(x) ((x).pgd)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) # define pgprot_val(x) ((x).pgprot)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) # define __pte(x) ((pte_t) { (x) } )
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194) # define __pmd(x) ((pmd_t) { (x) } )
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195) # define __pgprot(x) ((pgprot_t) { (x) } )
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197) #else /* !STRICT_MM_TYPECHECKS */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) * .. while these make it easier on the compiler
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201) # ifndef __ASSEMBLY__
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202) typedef unsigned long pte_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203) typedef unsigned long pmd_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204) typedef unsigned long pgd_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205) typedef unsigned long pgprot_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206) typedef struct page *pgtable_t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207) # endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) # define pte_val(x) (x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210) # define pmd_val(x) (x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) # define pgd_val(x) (x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) # define pgprot_val(x) (x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) # define __pte(x) (x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215) # define __pgd(x) (x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216) # define __pgprot(x) (x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) #endif /* !STRICT_MM_TYPECHECKS */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) #define PAGE_OFFSET RGN_BASE(RGN_KERNEL)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) #define VM_DATA_DEFAULT_FLAGS VM_DATA_FLAGS_TSK_EXEC
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223) #define GATE_ADDR RGN_BASE(RGN_GATE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226) * 0xa000000000000000+2*PERCPU_PAGE_SIZE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) * - 0xa000000000000000+3*PERCPU_PAGE_SIZE remain unmapped (guard page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) #define KERNEL_START (GATE_ADDR+__IA64_UL_CONST(0x100000000))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) #define PERCPU_ADDR (-PERCPU_PAGE_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) #define LOAD_OFFSET (KERNEL_START - KERNEL_TR_PAGE_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233) #define __HAVE_ARCH_GATE_AREA 1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235) #endif /* _ASM_IA64_PAGE_H */