Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 | // SPDX-License-Identifier: GPL-2.0 //! Tasks (threads and processes). //! //! C header: [`include/linux/sched.h`](srctree/include/linux/sched.h). use crate::{bindings, types::Opaque}; use core::{marker::PhantomData, ops::Deref, ptr}; /// Returns the currently running task. #[macro_export] macro_rules! current { () => { // SAFETY: Deref + addr-of below create a temporary `TaskRef` that cannot outlive the // caller. unsafe { &*$crate::task::Task::current() } }; } /// Wraps the kernel's `struct task_struct`. /// /// # Invariants /// /// All instances are valid tasks created by the C portion of the kernel. /// /// Instances of this type are always ref-counted, that is, a call to `get_task_struct` ensures /// that the allocation remains valid at least until the matching call to `put_task_struct`. /// /// # Examples /// /// The following is an example of getting the PID of the current thread with zero additional cost /// when compared to the C version: /// /// ``` /// let pid = current!().pid(); /// ``` /// /// Getting the PID of the current process, also zero additional cost: /// /// ``` /// let pid = current!().group_leader().pid(); /// ``` /// /// Getting the current task and storing it in some struct. The reference count is automatically /// incremented when creating `State` and decremented when it is dropped: /// /// ``` /// use kernel::{task::Task, types::ARef}; /// /// struct State { /// creator: ARef<Task>, /// index: u32, /// } /// /// impl State { /// fn new() -> Self { /// Self { /// creator: current!().into(), /// index: 0, /// } /// } /// } /// ``` #[repr(transparent)] pub struct Task(pub(crate) Opaque<bindings::task_struct>); // SAFETY: By design, the only way to access a `Task` is via the `current` function or via an // `ARef<Task>` obtained through the `AlwaysRefCounted` impl. This means that the only situation in // which a `Task` can be accessed mutably is when the refcount drops to zero and the destructor // runs. It is safe for that to happen on any thread, so it is ok for this type to be `Send`. unsafe impl Send for Task {} // SAFETY: It's OK to access `Task` through shared references from other threads because we're // either accessing properties that don't change (e.g., `pid`, `group_leader`) or that are properly // synchronised by C code (e.g., `signal_pending`). unsafe impl Sync for Task {} /// The type of process identifiers (PIDs). type Pid = bindings::pid_t; impl Task { /// Returns a task reference for the currently executing task/thread. /// /// The recommended way to get the current task/thread is to use the /// [`current`] macro because it is safe. /// /// # Safety /// /// Callers must ensure that the returned object doesn't outlive the current task/thread. pub unsafe fn current() -> impl Deref<Target = Task> { struct TaskRef<'a> { task: &'a Task, _not_send: PhantomData<*mut ()>, } impl Deref for TaskRef<'_> { type Target = Task; fn deref(&self) -> &Self::Target { self.task } } // SAFETY: Just an FFI call with no additional safety requirements. let ptr = unsafe { bindings::get_current() }; TaskRef { // SAFETY: If the current thread is still running, the current task is valid. Given // that `TaskRef` is not `Send`, we know it cannot be transferred to another thread // (where it could potentially outlive the caller). task: unsafe { &*ptr.cast() }, _not_send: PhantomData, } } /// Returns the group leader of the given task. pub fn group_leader(&self) -> &Task { // SAFETY: By the type invariant, we know that `self.0` is a valid task. Valid tasks always // have a valid group_leader. let ptr = unsafe { *ptr::addr_of!((*self.0.get()).group_leader) }; // SAFETY: The lifetime of the returned task reference is tied to the lifetime of `self`, // and given that a task has a reference to its group leader, we know it must be valid for // the lifetime of the returned task reference. unsafe { &*ptr.cast() } } /// Returns the PID of the given task. pub fn pid(&self) -> Pid { // SAFETY: By the type invariant, we know that `self.0` is a valid task. Valid tasks always // have a valid pid. unsafe { *ptr::addr_of!((*self.0.get()).pid) } } /// Determines whether the given task has pending signals. pub fn signal_pending(&self) -> bool { // SAFETY: By the type invariant, we know that `self.0` is valid. unsafe { bindings::signal_pending(self.0.get()) != 0 } } /// Wakes up the task. pub fn wake_up(&self) { // SAFETY: By the type invariant, we know that `self.0.get()` is non-null and valid. // And `wake_up_process` is safe to be called for any valid task, even if the task is // running. unsafe { bindings::wake_up_process(self.0.get()) }; } } // SAFETY: The type invariants guarantee that `Task` is always ref-counted. unsafe impl crate::types::AlwaysRefCounted for Task { fn inc_ref(&self) { // SAFETY: The existence of a shared reference means that the refcount is nonzero. unsafe { bindings::get_task_struct(self.0.get()) }; } unsafe fn dec_ref(obj: ptr::NonNull<Self>) { // SAFETY: The safety requirements guarantee that the refcount is nonzero. unsafe { bindings::put_task_struct(obj.cast().as_ptr()) } } } |