summaryrefslogtreecommitdiff
path: root/mm/memory.c
diff options
context:
space:
mode:
authorLinus Torvalds <torvalds@linux-foundation.org>2023-06-16 02:17:48 +0300
committerLinus Torvalds <torvalds@linux-foundation.org>2023-06-25 00:12:57 +0300
commiteda0047296a16d65a7f2bc60a408f70d178b2014 (patch)
tree934ab40aff1cafd4ca1da8a2c5ead6aba7adabe1 /mm/memory.c
parentc2508ec5a58db67093f4fb8bf89a9a7c53a109e9 (diff)
downloadlinux-eda0047296a16d65a7f2bc60a408f70d178b2014.tar.xz
mm: make the page fault mmap locking killable
This is done as a separate patch from introducing the new lock_mm_and_find_vma() helper, because while it's an obvious change, it's not what x86 used to do in this area. We already abort the page fault on fatal signals anyway, so why should we wait for the mmap lock only to then abort later? With the new helper function that returns without the lock held on failure anyway, this is particularly easy and straightforward. Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm/memory.c')
-rw-r--r--mm/memory.c6
1 files changed, 2 insertions, 4 deletions
diff --git a/mm/memory.c b/mm/memory.c
index 1a427097b71f..1dff248805bf 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -5279,8 +5279,7 @@ static inline bool get_mmap_lock_carefully(struct mm_struct *mm, struct pt_regs
return false;
}
- mmap_read_lock(mm);
- return true;
+ return !mmap_read_lock_killable(mm);
}
static inline bool mmap_upgrade_trylock(struct mm_struct *mm)
@@ -5304,8 +5303,7 @@ static inline bool upgrade_mmap_lock_carefully(struct mm_struct *mm, struct pt_r
if (!search_exception_tables(ip))
return false;
}
- mmap_write_lock(mm);
- return true;
+ return !mmap_write_lock_killable(mm);
}
/*