1/* SPDX-License-Identifier: GPL-2.0 */
2#ifndef _ASM_X86_PAGE_64_H
3#define _ASM_X86_PAGE_64_H
4
5#include <asm/page_64_types.h>
6
7#ifndef __ASSEMBLER__
8#include <asm/cpufeatures.h>
9#include <asm/alternative.h>
10
11#include <linux/kmsan-checks.h>
12
13/* duplicated to the one in bootmem.h */
14extern unsigned long max_pfn;
15extern unsigned long phys_base;
16
17extern unsigned long page_offset_base;
18extern unsigned long vmalloc_base;
19extern unsigned long vmemmap_base;
20extern unsigned long direct_map_physmem_end;
21
22static __always_inline unsigned long __phys_addr_nodebug(unsigned long x)
23{
24 unsigned long y = x - __START_KERNEL_map;
25
26 /* use the carry flag to determine if x was < __START_KERNEL_map */
27 x = y + ((x > y) ? phys_base : (__START_KERNEL_map - PAGE_OFFSET));
28
29 return x;
30}
31
32#ifdef CONFIG_DEBUG_VIRTUAL
33extern unsigned long __phys_addr(unsigned long);
34extern unsigned long __phys_addr_symbol(unsigned long);
35#else
36#define __phys_addr(x) __phys_addr_nodebug(x)
37#define __phys_addr_symbol(x) \
38 ((unsigned long)(x) - __START_KERNEL_map + phys_base)
39#endif
40
41#define __phys_reloc_hide(x) (x)
42
43void clear_page_orig(void *page);
44void clear_page_rep(void *page);
45void clear_page_erms(void *page);
46
47static inline void clear_page(void *page)
48{
49 /*
50 * Clean up KMSAN metadata for the page being cleared. The assembly call
51 * below clobbers @page, so we perform unpoisoning before it.
52 */
53 kmsan_unpoison_memory(address: page, PAGE_SIZE);
54 alternative_call_2(clear_page_orig,
55 clear_page_rep, X86_FEATURE_REP_GOOD,
56 clear_page_erms, X86_FEATURE_ERMS,
57 "=D" (page),
58 "D" (page),
59 "cc", "memory", "rax", "rcx");
60}
61
62void copy_page(void *to, void *from);
63KCFI_REFERENCE(copy_page);
64
65/*
66 * User space process size. This is the first address outside the user range.
67 * There are a few constraints that determine this:
68 *
69 * On Intel CPUs, if a SYSCALL instruction is at the highest canonical
70 * address, then that syscall will enter the kernel with a
71 * non-canonical return address, and SYSRET will explode dangerously.
72 * We avoid this particular problem by preventing anything
73 * from being mapped at the maximum canonical address.
74 *
75 * On AMD CPUs in the Ryzen family, there's a nasty bug in which the
76 * CPUs malfunction if they execute code from the highest canonical page.
77 * They'll speculate right off the end of the canonical space, and
78 * bad things happen. This is worked around in the same way as the
79 * Intel problem.
80 *
81 * With page table isolation enabled, we map the LDT in ... [stay tuned]
82 */
83static __always_inline unsigned long task_size_max(void)
84{
85 unsigned long ret;
86
87 alternative_io("movq %[small],%0","movq %[large],%0",
88 X86_FEATURE_LA57,
89 "=r" (ret),
90 [small] "i" ((1ul << 47)-PAGE_SIZE),
91 [large] "i" ((1ul << 56)-PAGE_SIZE));
92
93 return ret;
94}
95
96#endif /* !__ASSEMBLER__ */
97
98#ifdef CONFIG_X86_VSYSCALL_EMULATION
99# define __HAVE_ARCH_GATE_AREA 1
100#endif
101
102#endif /* _ASM_X86_PAGE_64_H */
103

Provided by KDAB

Privacy Policy
Improve your Profiling and Debugging skills
Find out more

source code of linux/arch/x86/include/asm/page_64.h