summaryrefslogtreecommitdiff
path: root/arch/x86/um/shared/sysdep/faultinfo_32.h
diff options
context:
space:
mode:
authorJohannes Berg <johannes.berg@intel.com>2025-04-04 17:05:19 +0200
committerJohannes Berg <johannes.berg@intel.com>2025-05-05 10:06:51 +0200
commit68025adfc13e6cd15eebe2293f77659f47daf13b (patch)
treeed250635fcbce5f1c21a12af1785d7e3891ec860 /arch/x86/um/shared/sysdep/faultinfo_32.h
parent92a09c47464d040866cf2b4cd052bc60555185fb (diff)
um: fix _nofault accesses
Nathan reported [1] that when built with clang, the um kernel crashes pretty much immediately. This turned out to be an issue with the inline assembly I had added, when clang used %rax/%eax for both operands. Reorder it so current->thread.segv_continue is written first, and then the lifetime of _faulted won't have overlap with the lifetime of segv_continue. In the email thread Benjamin also pointed out that current->mm is only NULL for true kernel tasks, but we could do this for a userspace task, so the current->thread.segv_continue logic must be lifted out of the mm==NULL check. Finally, while looking at this, put a barrier() so the NULL assignment to thread.segv_continue cannot be reorder before the possibly faulting operation. Reported-by: Nathan Chancellor <nathan@kernel.org> Closes: https://lore.kernel.org/r/20250402221254.GA384@ax162 [1] Fixes: d1d7f01f7cd3 ("um: mark rodata read-only and implement _nofault accesses") Tested-by: Nathan Chancellor <nathan@kernel.org> Signed-off-by: Johannes Berg <johannes.berg@intel.com>
Diffstat (limited to 'arch/x86/um/shared/sysdep/faultinfo_32.h')
-rw-r--r--arch/x86/um/shared/sysdep/faultinfo_32.h2
1 files changed, 1 insertions, 1 deletions
diff --git a/arch/x86/um/shared/sysdep/faultinfo_32.h b/arch/x86/um/shared/sysdep/faultinfo_32.h
index ab5c8e47049c..9193a7790a71 100644
--- a/arch/x86/um/shared/sysdep/faultinfo_32.h
+++ b/arch/x86/um/shared/sysdep/faultinfo_32.h
@@ -31,8 +31,8 @@ struct faultinfo {
#define ___backtrack_faulted(_faulted) \
asm volatile ( \
- "mov $0, %0\n" \
"movl $__get_kernel_nofault_faulted_%=,%1\n" \
+ "mov $0, %0\n" \
"jmp _end_%=\n" \
"__get_kernel_nofault_faulted_%=:\n" \
"mov $1, %0;" \