// SPDX-License-Identifier: GPL-2.0 // Copyright (C) 2024 Google LLC. //! Memory management. //! //! This module deals with managing the address space of userspace processes. Each process has an //! instance of [`Mm`], which keeps track of multiple VMAs (virtual memory areas). Each VMA //! corresponds to a region of memory that the userspace process can access, and the VMA lets you //! control what happens when userspace reads or writes to that region of memory. //! //! C header: [`include/linux/mm.h`](srctree/include/linux/mm.h) #![cfg(CONFIG_MMU)] use crate::{ bindings, types::{ARef, AlwaysRefCounted, NotThreadSafe, Opaque}, }; use core::{ops::Deref, ptr::NonNull}; pub mod virt; use virt::VmaRef; /// A wrapper for the kernel's `struct mm_struct`. /// /// This represents the address space of a userspace process, so each process has one `Mm` /// instance. It may hold many VMAs internally. /// /// There is a counter called `mm_users` that counts the users of the address space; this includes /// the userspace process itself, but can also include kernel threads accessing the address space. /// Once `mm_users` reaches zero, this indicates that the address space can be destroyed. To access /// the address space, you must prevent `mm_users` from reaching zero while you are accessing it. /// The [`MmWithUser`] type represents an address space where this is guaranteed, and you can /// create one using [`mmget_not_zero`]. /// /// The `ARef` smart pointer holds an `mmgrab` refcount. Its destructor may sleep. /// /// # Invariants /// /// Values of this type are always refcounted using `mmgrab`. /// /// [`mmget_not_zero`]: Mm::mmget_not_zero #[repr(transparent)] pub struct Mm { mm: Opaque, } // SAFETY: It is safe to call `mmdrop` on another thread than where `mmgrab` was called. unsafe impl Send for Mm {} // SAFETY: All methods on `Mm` can be called in parallel from several threads. unsafe impl Sync for Mm {} // SAFETY: By the type invariants, this type is always refcounted. unsafe impl AlwaysRefCounted for Mm { #[inline] fn inc_ref(&self) { // SAFETY: The pointer is valid since self is a reference. unsafe { bindings::mmgrab(self.as_raw()) }; } #[inline] unsafe fn dec_ref(obj: NonNull) { // SAFETY: The caller is giving up their refcount. unsafe { bindings::mmdrop(obj.cast().as_ptr()) }; } } /// A wrapper for the kernel's `struct mm_struct`. /// /// This type is like [`Mm`], but with non-zero `mm_users`. It can only be used when `mm_users` can /// be proven to be non-zero at compile-time, usually because the relevant code holds an `mmget` /// refcount. It can be used to access the associated address space. /// /// The `ARef` smart pointer holds an `mmget` refcount. Its destructor may sleep. /// /// # Invariants /// /// Values of this type are always refcounted using `mmget`. The value of `mm_users` is non-zero. #[repr(transparent)] pub struct MmWithUser { mm: Mm, } // SAFETY: It is safe to call `mmput` on another thread than where `mmget` was called. unsafe impl Send for MmWithUser {} // SAFETY: All methods on `MmWithUser` can be called in parallel from several threads. unsafe impl Sync for MmWithUser {} // SAFETY: By the type invariants, this type is always refcounted. unsafe impl AlwaysRefCounted for MmWithUser { #[inline] fn inc_ref(&self) { // SAFETY: The pointer is valid since self is a reference. unsafe { bindings::mmget(self.as_raw()) }; } #[inline] unsafe fn dec_ref(obj: NonNull) { // SAFETY: The caller is giving up their refcount. unsafe { bindings::mmput(obj.cast().as_ptr()) }; } } // Make all `Mm` methods available on `MmWithUser`. impl Deref for MmWithUser { type Target = Mm; #[inline] fn deref(&self) -> &Mm { &self.mm } } /// A wrapper for the kernel's `struct mm_struct`. /// /// This type is identical to `MmWithUser` except that it uses `mmput_async` when dropping a /// refcount. This means that the destructor of `ARef` is safe to call in atomic /// context. /// /// # Invariants /// /// Values of this type are always refcounted using `mmget`. The value of `mm_users` is non-zero. #[repr(transparent)] pub struct MmWithUserAsync { mm: MmWithUser, } // SAFETY: It is safe to call `mmput_async` on another thread than where `mmget` was called. unsafe impl Send for MmWithUserAsync {} // SAFETY: All methods on `MmWithUserAsync` can be called in parallel from several threads. unsafe impl Sync for MmWithUserAsync {} // SAFETY: By the type invariants, this type is always refcounted. unsafe impl AlwaysRefCounted for MmWithUserAsync { #[inline] fn inc_ref(&self) { // SAFETY: The pointer is valid since self is a reference. unsafe { bindings::mmget(self.as_raw()) }; } #[inline] unsafe fn dec_ref(obj: NonNull) { // SAFETY: The caller is giving up their refcount. unsafe { bindings::mmput_async(obj.cast().as_ptr()) }; } } // Make all `MmWithUser` methods available on `MmWithUserAsync`. impl Deref for MmWithUserAsync { type Target = MmWithUser; #[inline] fn deref(&self) -> &MmWithUser { &self.mm } } // These methods are safe to call even if `mm_users` is zero. impl Mm { /// Returns a raw pointer to the inner `mm_struct`. #[inline] pub fn as_raw(&self) -> *mut bindings::mm_struct { self.mm.get() } /// Obtain a reference from a raw pointer. /// /// # Safety /// /// The caller must ensure that `ptr` points at an `mm_struct`, and that it is not deallocated /// during the lifetime 'a. #[inline] pub unsafe fn from_raw<'a>(ptr: *const bindings::mm_struct) -> &'a Mm { // SAFETY: Caller promises that the pointer is valid for 'a. Layouts are compatible due to // repr(transparent). unsafe { &*ptr.cast() } } /// Calls `mmget_not_zero` and returns a handle if it succeeds. #[inline] pub fn mmget_not_zero(&self) -> Option> { // SAFETY: The pointer is valid since self is a reference. let success = unsafe { bindings::mmget_not_zero(self.as_raw()) }; if success { // SAFETY: We just created an `mmget` refcount. Some(unsafe { ARef::from_raw(NonNull::new_unchecked(self.as_raw().cast())) }) } else { None } } } // These methods require `mm_users` to be non-zero. impl MmWithUser { /// Obtain a reference from a raw pointer. /// /// # Safety /// /// The caller must ensure that `ptr` points at an `mm_struct`, and that `mm_users` remains /// non-zero for the duration of the lifetime 'a. #[inline] pub unsafe fn from_raw<'a>(ptr: *const bindings::mm_struct) -> &'a MmWithUser { // SAFETY: Caller promises that the pointer is valid for 'a. The layout is compatible due // to repr(transparent). unsafe { &*ptr.cast() } } /// Use `mmput_async` when dropping this refcount. #[inline] pub fn into_mmput_async(me: ARef) -> ARef { // SAFETY: The layouts and invariants are compatible. unsafe { ARef::from_raw(ARef::into_raw(me).cast()) } } /// Attempt to access a vma using the vma read lock. /// /// This is an optimistic trylock operation, so it may fail if there is contention. In that /// case, you should fall back to taking the mmap read lock. /// /// When per-vma locks are disabled, this always returns `None`. #[inline] pub fn lock_vma_under_rcu(&self, vma_addr: usize) -> Option> { #[cfg(CONFIG_PER_VMA_LOCK)] { // SAFETY: Calling `bindings::lock_vma_under_rcu` is always okay given an mm where // `mm_users` is non-zero. let vma = unsafe { bindings::lock_vma_under_rcu(self.as_raw(), vma_addr) }; if !vma.is_null() { return Some(VmaReadGuard { // SAFETY: If `lock_vma_under_rcu` returns a non-null ptr, then it points at a // valid vma. The vma is stable for as long as the vma read lock is held. vma: unsafe { VmaRef::from_raw(vma) }, _nts: NotThreadSafe, }); } } // Silence warnings about unused variables. #[cfg(not(CONFIG_PER_VMA_LOCK))] let _ = vma_addr; None } /// Lock the mmap read lock. #[inline] pub fn mmap_read_lock(&self) -> MmapReadGuard<'_> { // SAFETY: The pointer is valid since self is a reference. unsafe { bindings::mmap_read_lock(self.as_raw()) }; // INVARIANT: We just acquired the read lock. MmapReadGuard { mm: self, _nts: NotThreadSafe, } } /// Try to lock the mmap read lock. #[inline] pub fn mmap_read_trylock(&self) -> Option> { // SAFETY: The pointer is valid since self is a reference. let success = unsafe { bindings::mmap_read_trylock(self.as_raw()) }; if success { // INVARIANT: We just acquired the read lock. Some(MmapReadGuard { mm: self, _nts: NotThreadSafe, }) } else { None } } } /// A guard for the mmap read lock. /// /// # Invariants /// /// This `MmapReadGuard` guard owns the mmap read lock. pub struct MmapReadGuard<'a> { mm: &'a MmWithUser, // `mmap_read_lock` and `mmap_read_unlock` must be called on the same thread _nts: NotThreadSafe, } impl<'a> MmapReadGuard<'a> { /// Look up a vma at the given address. #[inline] pub fn vma_lookup(&self, vma_addr: usize) -> Option<&virt::VmaRef> { // SAFETY: By the type invariants we hold the mmap read guard, so we can safely call this // method. Any value is okay for `vma_addr`. let vma = unsafe { bindings::vma_lookup(self.mm.as_raw(), vma_addr) }; if vma.is_null() { None } else { // SAFETY: We just checked that a vma was found, so the pointer references a valid vma. // // Furthermore, the returned vma is still under the protection of the read lock guard // and can be used while the mmap read lock is still held. That the vma is not used // after the MmapReadGuard gets dropped is enforced by the borrow-checker. unsafe { Some(virt::VmaRef::from_raw(vma)) } } } } impl Drop for MmapReadGuard<'_> { #[inline] fn drop(&mut self) { // SAFETY: We hold the read lock by the type invariants. unsafe { bindings::mmap_read_unlock(self.mm.as_raw()) }; } } /// A guard for the vma read lock. /// /// # Invariants /// /// This `VmaReadGuard` guard owns the vma read lock. pub struct VmaReadGuard<'a> { vma: &'a VmaRef, // `vma_end_read` must be called on the same thread as where the lock was taken _nts: NotThreadSafe, } // Make all `VmaRef` methods available on `VmaReadGuard`. impl Deref for VmaReadGuard<'_> { type Target = VmaRef; #[inline] fn deref(&self) -> &VmaRef { self.vma } } impl Drop for VmaReadGuard<'_> { #[inline] fn drop(&mut self) { // SAFETY: We hold the read lock by the type invariants. unsafe { bindings::vma_end_read(self.vma.as_ptr()) }; } }