//! The rusty version of iothreads from the cpp code, to be consumed by native rust code. This isn't //! ported directly from the cpp code so we can use rust threads instead of using pthreads. use crate::flog::{FloggableDebug, FLOG}; use once_cell::race::OnceBox; use std::num::NonZeroU64; use std::sync::atomic::{AtomicBool, Ordering}; use std::sync::{Arc, Mutex}; use std::thread::{self, ThreadId}; use std::time::{Duration, Instant}; impl FloggableDebug for ThreadId {} // We don't want to use a full-blown Lazy for the cached main thread id, but we can't use // AtomicU64 since std::thread::ThreadId::as_u64() is a nightly-only feature (issue #67939, // thread_id_value). We also can't safely transmute `ThreadId` to `NonZeroU64` because there's no // guarantee that's what the underlying type will always be on all platforms and in all cases, // `ThreadId` isn't marked `#[repr(transparent)]`. We could generate our own thread-local value, but // `#[thread_local]` is nightly-only while the stable `thread_local!()` macro doesn't generate // efficient/fast/low-overhead code. /// The thread id of the main thread, as set by [`init()`] at startup. static mut MAIN_THREAD_ID: Option = None; /// Used to bypass thread assertions when testing. static THREAD_ASSERTS_CFG_FOR_TESTING: AtomicBool = AtomicBool::new(false); /// This allows us to notice when we've forked. static IS_FORKED_PROC: AtomicBool = AtomicBool::new(false); /// Maximum number of threads for the IO thread pool. const IO_MAX_THREADS: usize = 1024; /// How long an idle [`ThreadPool`] thread will wait for work (against the condition variable) /// before exiting. const IO_WAIT_FOR_WORK_DURATION: Duration = Duration::from_millis(500); /// The iothreads [`ThreadPool`] singleton. Used to lift I/O off of the main thread and used for /// completions, etc. static IO_THREAD_POOL: OnceBox> = OnceBox::new(); /// The event signaller singleton used for completions and queued main thread requests. static NOTIFY_SIGNALLER: once_cell::sync::Lazy<&'static crate::fd_monitor::FdEventSignaller> = once_cell::sync::Lazy::new(|| unsafe { // This is leaked to avoid C++-side destructors. When ported fully to rust, we won't need to // leak anything. let signaller = crate::fd_monitor::new_fd_event_signaller(); let signaller_ref: &crate::fd_monitor::FdEventSignaller = signaller.as_ref().unwrap(); let result = std::mem::transmute(signaller_ref); std::mem::forget(signaller); result }); #[cxx::bridge] mod ffi { unsafe extern "C++" { include!("callback.h"); #[rust_name = "CppCallback"] type callback_t; fn invoke(&self) -> *const u8; fn invoke_with_param(&self, param: *const u8) -> *const u8; } extern "Rust" { #[cxx_name = "ASSERT_IS_MAIN_THREAD"] fn assert_is_main_thread(); #[cxx_name = "ASSERT_IS_BACKGROUND_THREAD"] fn assert_is_background_thread(); #[cxx_name = "ASSERT_IS_NOT_FORKED_CHILD"] fn assert_is_not_forked_child(); fn configure_thread_assertions_for_testing(); fn is_main_thread() -> bool; fn is_forked_child() -> bool; } extern "Rust" { #[cxx_name = "make_detached_pthread"] fn spawn_ffi(callback: &SharedPtr) -> bool; fn asan_before_exit(); fn asan_maybe_exit(code: i32); } extern "Rust" { fn iothread_port() -> i32; fn iothread_service_main(); #[cxx_name = "iothread_service_main_with_timeout"] fn iothread_service_main_with_timeout_ffi(timeout_usec: u64); #[cxx_name = "iothread_drain_all"] fn iothread_drain_all_ffi(); #[cxx_name = "iothread_perform"] fn iothread_perform_ffi(callback: &SharedPtr); #[cxx_name = "iothread_perform_cantwait"] fn iothread_perform_cant_wait_ffi(callback: &SharedPtr); } extern "Rust" { #[cxx_name = "debounce_t"] type Debounce; #[cxx_name = "perform"] fn perform_ffi(&self, callback: &SharedPtr) -> u64; #[cxx_name = "perform_with_completion"] fn perform_with_completion_ffi( &self, callback: &SharedPtr, completion: &SharedPtr, ) -> u64; #[cxx_name = "new_debounce_t"] fn new_debounce_ffi(timeout_ms: u64) -> Box; } } unsafe impl Send for ffi::CppCallback {} unsafe impl Sync for ffi::CppCallback {} fn iothread_service_main_with_timeout_ffi(timeout_usec: u64) { iothread_service_main_with_timeout(Duration::from_micros(timeout_usec)) } fn iothread_drain_all_ffi() { unsafe { iothread_drain_all() } } fn iothread_perform_ffi(callback: &cxx::SharedPtr) { let callback = callback.clone(); iothread_perform(move || { callback.invoke(); }); } fn iothread_perform_cant_wait_ffi(callback: &cxx::SharedPtr) { let callback = callback.clone(); iothread_perform_cant_wait(move || { callback.invoke(); }); } /// A [`ThreadPool`] or [`Debounce`] work request. type WorkItem = Box; /// The queue of [`WorkItem`]s to be executed on the main thread. This is added to from /// [`Debounce::enqueue_main_thread()`] and read from in `iothread_service_main()`. /// /// Since items are enqueued from various background threads then read by the main thread, the work /// items must implement `Send`. static MAIN_THREAD_QUEUE: Mutex> = Mutex::new(Vec::new()); /// Initialize some global static variables. Must be called at startup from the main thread. pub fn init() { unsafe { if MAIN_THREAD_ID.is_some() { panic!("threads::init() must only be called once (at startup)!"); } MAIN_THREAD_ID = Some(thread::current().id()); } extern "C" fn child_post_fork() { IS_FORKED_PROC.store(true, Ordering::Relaxed); } unsafe { let result = libc::pthread_atfork(None, None, Some(child_post_fork)); assert_eq!(result, 0, "pthread_atfork() failure: {}", errno::errno()); } IO_THREAD_POOL .set(Box::new(Mutex::new(ThreadPool::new(1, IO_MAX_THREADS)))) .expect("IO_THREAD_POOL has already been initialized!"); } #[inline(always)] fn main_thread_id() -> ThreadId { #[cold] fn init_not_called() -> ! { panic!("threads::init() was not called at startup!"); } match unsafe { MAIN_THREAD_ID } { None => init_not_called(), Some(id) => id, } } #[inline(always)] pub fn is_main_thread() -> bool { thread::current().id() == main_thread_id() } #[inline(always)] pub fn assert_is_main_thread() { #[cold] fn not_main_thread() -> ! { panic!("Function is not running on the main thread!"); } if !is_main_thread() && !THREAD_ASSERTS_CFG_FOR_TESTING.load(Ordering::Relaxed) { not_main_thread(); } } #[inline(always)] pub fn assert_is_background_thread() { #[cold] fn not_background_thread() -> ! { panic!("Function is not allowed to be called on the main thread!"); } if is_main_thread() && !THREAD_ASSERTS_CFG_FOR_TESTING.load(Ordering::Relaxed) { not_background_thread(); } } pub fn configure_thread_assertions_for_testing() { THREAD_ASSERTS_CFG_FOR_TESTING.store(true, Ordering::Relaxed); } pub fn is_forked_child() -> bool { IS_FORKED_PROC.load(Ordering::Relaxed) } #[inline(always)] pub fn assert_is_not_forked_child() { #[cold] fn panic_is_forked_child() { panic!("Function called from forked child!"); } if is_forked_child() { panic_is_forked_child(); } } /// The rusty version of `iothreads::make_detached_pthread()`. We will probably need a /// `spawn_scoped` version of the same to handle some more advanced borrow cases safely, and maybe /// an unsafe version that doesn't do any lifetime checking akin to /// `spawn_unchecked()`[std::thread::Builder::spawn_unchecked], which is a nightly-only feature. /// /// Returns a boolean indicating whether or not the thread was successfully launched. Failure here /// is not dependent on the passed callback and implies a system error (likely insufficient /// resources). pub fn spawn(callback: F) -> bool { // The spawned thread inherits our signal mask. Temporarily block signals, spawn the thread, and // then restore it. But we must not block SIGBUS, SIGFPE, SIGILL, or SIGSEGV; that's undefined // (#7837). Conservatively don't try to mask SIGKILL or SIGSTOP either; that's ignored on Linux // but maybe has an effect elsewhere. let saved_set = unsafe { let mut new_set: libc::sigset_t = std::mem::zeroed(); let new_set = &mut new_set as *mut _; libc::sigfillset(new_set); libc::sigdelset(new_set, libc::SIGILL); // bad jump libc::sigdelset(new_set, libc::SIGFPE); // divide-by-zero libc::sigdelset(new_set, libc::SIGBUS); // unaligned memory access libc::sigdelset(new_set, libc::SIGSEGV); // bad memory access libc::sigdelset(new_set, libc::SIGSTOP); // unblockable libc::sigdelset(new_set, libc::SIGKILL); // unblockable let mut saved_set: libc::sigset_t = std::mem::zeroed(); let result = libc::pthread_sigmask(libc::SIG_BLOCK, new_set, &mut saved_set as *mut _); assert_eq!(result, 0, "Failed to override thread signal mask!"); saved_set }; // Spawn a thread. If this fails, it means there's already a bunch of threads; it is very // unlikely that they are all on the verge of exiting, so one is likely to be ready to handle // extant requests. So we can ignore failure with some confidence. // We don't have to port the PTHREAD_CREATE_DETACHED logic. Rust threads are detached // automatically if the returned join handle is dropped. let result = match std::thread::Builder::new().spawn(move || { (callback)(); }) { Ok(handle) => { let thread_id = handle.thread().id(); FLOG!(iothread, "rust thread", thread_id, "spawned"); // Drop the handle to detach the thread drop(handle); true } Err(e) => { eprintln!("rust thread spawn failure: {e}"); false } }; // Restore our sigmask unsafe { let result = libc::pthread_sigmask( libc::SIG_SETMASK, &saved_set as *const _, std::ptr::null_mut(), ); assert_eq!(result, 0, "Failed to restore thread signal mask!"); }; result } fn spawn_ffi(callback: &cxx::SharedPtr) -> bool { let callback = callback.clone(); spawn(move || { callback.invoke(); }) } /// Exits calling onexit handlers if running under ASAN, otherwise does nothing. /// /// This function is always defined but is a no-op if not running under ASAN. This is to make it /// more ergonomic to call it in general and also makes it possible to call it via ffi at all. pub fn asan_maybe_exit(code: i32) { if cfg!(feature = "asan") { asan_before_exit(); unsafe { libc::exit(code); } } } /// When running under ASAN, free up some allocations that would normally have been left for the OS /// to reclaim to avoid some false positive LSAN reports. /// /// This function is always defined but is a no-op if not running under ASAN. This is to make it /// more ergonomic to call it in general and also makes it possible to call it via ffi at all. pub fn asan_before_exit() { if cfg!(feature = "asan") && !is_forked_child() { // Free ncurses terminal state crate::curses::reset(); } } /// Data shared between the thread pool [`ThreadPool`] and worker threads [`WorkerThread`]. #[derive(Default)] struct ThreadPoolProtected { /// The queue of outstanding, unclaimed work requests pub request_queue: std::collections::VecDeque, /// The number of threads that exist in the pool pub total_threads: usize, /// The number of threads waiting for more work (i.e. idle threads) pub waiting_threads: usize, } /// Data behind an [`Arc`] to share between the [`ThreadPool`] and [`WorkerThread`] instances. #[derive(Default)] struct ThreadPoolShared { /// The mutex to access shared state between [`ThreadPool`] and [`WorkerThread`] instances. This /// is accessed both standalone and via [`cond_var`](Self::cond_var). mutex: Mutex, /// The condition variable used to wake up waiting threads. This is tied to [`mutex`](Self::mutex). cond_var: std::sync::Condvar, } pub struct ThreadPool { /// The data which needs to be shared with worker threads. shared: Arc, /// The minimum number of threads that will be kept waiting even when idle in the pool. soft_min_threads: usize, /// The maximum number of threads that will be created to service outstanding work requests, by /// default. This may be bypassed. max_threads: usize, } impl std::fmt::Debug for ThreadPool { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { f.debug_struct("ThreadPool") .field("min_threads", &self.soft_min_threads) .field("max_threads", &self.max_threads) .finish() } } impl ThreadPool { /// Construct a new `ThreadPool` instance with the specified min and max num of threads. pub fn new(soft_min_threads: usize, max_threads: usize) -> Self { ThreadPool { shared: Default::default(), soft_min_threads, max_threads, } } /// Enqueue a new work item onto the thread pool. /// /// The function `func` will execute on one of the pool's background threads. If `cant_wait` is /// set, the thread limit may be disregarded if extant threads are busy. /// /// Returns the number of threads that were alive when the work item was enqueued. pub fn perform(&mut self, func: F, cant_wait: bool) -> usize { let work_item = Box::new(func); self.perform_inner(work_item, cant_wait) } fn perform_inner(&mut self, f: WorkItem, cant_wait: bool) -> usize { enum ThreadAction { None, Wake, Spawn, } let local_thread_count; let thread_action = { let mut data = self.shared.mutex.lock().expect("Mutex poisoned!"); local_thread_count = data.total_threads; data.request_queue.push_back(f); FLOG!( iothread, "enqueuing work item (count is ", data.request_queue.len(), ")" ); if data.waiting_threads >= data.request_queue.len() { // There are enough waiting threads, wake one up. ThreadAction::Wake } else if cant_wait || data.total_threads < self.max_threads { // No threads are idle waiting but we can or must spawn a new thread to service the // request. data.total_threads += 1; ThreadAction::Spawn } else { // There is no need to do anything because we've reached the max number of threads. ThreadAction::None } }; // Act only after unlocking the mutex. match thread_action { ThreadAction::None => (), ThreadAction::Wake => { // Wake a thread if we decided to do so. FLOG!(iothread, "notifying thread ", std::thread::current().id()); self.shared.cond_var.notify_one(); } ThreadAction::Spawn => { // Spawn a thread. If this fails, it means there are already a bunch of worker // threads and it is very unlikely that they are all about to exit so one is likely // able to handle the incoming request. This means we can ignore the failure with // some degree of confidence. (This is also not an error we expect to routinely run // into under normal, non-resource-starved circumstances.) if self.spawn_thread() { FLOG!(iothread, "pthread spawned"); } else { // We failed to spawn a thread; decrement the thread count. self.shared .mutex .lock() .expect("Mutex poisoned!") .total_threads -= 1; } } } local_thread_count } /// Attempt to spawn a new worker thread. fn spawn_thread(&mut self) -> bool { let shared = Arc::clone(&self.shared); let soft_min_threads = self.soft_min_threads; self::spawn(move || { let worker = WorkerThread { shared, soft_min_threads, }; worker.run(); }) } } pub struct WorkerThread { /// The data shared with the [`ThreadPool`]. shared: Arc, /// The soft min number of threads for the associated [`ThreadPool`]. soft_min_threads: usize, } impl WorkerThread { /// The worker loop entry point for this thread. fn run(mut self) { while let Some(work_item) = self.dequeue_work_or_commit_to_exit() { FLOG!( iothread, "pthread ", std::thread::current().id(), " got work" ); // Perform the work work_item(); } FLOG!( iothread, "pthread ", std::thread::current().id(), " exiting" ); } /// Dequeue a work item (perhaps waiting on the condition variable) or commit to exiting by /// reducing the active thread count. fn dequeue_work_or_commit_to_exit(&mut self) -> Option { let mut data = self.shared.mutex.lock().expect("Mutex poisoned!"); // If the queue is empty, check to see if we should wait. We should wait if our exiting // would drop us below our soft thread count minimum. if data.request_queue.is_empty() && data.total_threads == self.soft_min_threads && IO_WAIT_FOR_WORK_DURATION > Duration::ZERO { data.waiting_threads += 1; data = self .shared .cond_var .wait_timeout(data, IO_WAIT_FOR_WORK_DURATION) .expect("Mutex poisoned!") .0; data.waiting_threads -= 1; } // Now that we've (perhaps) waited, see if there's something on the queue. let result = data.request_queue.pop_front(); // If we are returning None then ensure we balance the thread count increment from when we // were created. This has to be done here in this awkward place because we've already // committed to exiting - we will never pick up more work. So we need to make sure to // decrement the thread count while holding the lock as we have effectively already exited. if result.is_none() { data.total_threads -= 1; } return result; } } /// Returns a [`MutexGuard`](std::sync::MutexGuard) containing the IO [`ThreadPool`]. fn borrow_io_thread_pool() -> std::sync::MutexGuard<'static, ThreadPool> { IO_THREAD_POOL .get() .unwrap() .lock() .expect("Mutex poisoned!") } /// Enqueues work on the IO thread pool singleton. pub fn iothread_perform(f: impl FnOnce() + 'static + Send) { let mut thread_pool = borrow_io_thread_pool(); thread_pool.perform(f, false); } /// Enqueues priority work on the IO thread pool singleton, disregarding the thread limit. /// /// It does its best to spawn a thread if all other threads are occupied. This is primarily for /// cases where deferring creation of a new thread might lead to a deadlock. pub fn iothread_perform_cant_wait(f: impl FnOnce() + 'static + Send) { let mut thread_pool = borrow_io_thread_pool(); thread_pool.perform(f, true); } pub fn iothread_port() -> i32 { i32::from(NOTIFY_SIGNALLER.read_fd()) } pub fn iothread_service_main_with_timeout(timeout: Duration) { if crate::fd_readable_set::is_fd_readable(iothread_port(), timeout.as_millis() as u64) { iothread_service_main(); } } pub fn iothread_service_main() { self::assert_is_main_thread(); // Note: the order here is important. We must consume events before handling requests, as // posting uses the opposite order. NOTIFY_SIGNALLER.try_consume(); // Move the queue to a local variable. The MAIN_THREAD_QUEUE lock is not held after this. let queue = std::mem::take(&mut *MAIN_THREAD_QUEUE.lock().expect("Mutex poisoned!")); // Perform each completion in order. for func in queue { (func)(); } } /// Does nasty polling via select() and marked as unsafe because it should only be used for testing. pub unsafe fn iothread_drain_all() { while borrow_io_thread_pool() .shared .mutex .lock() .expect("Mutex poisoned!") .total_threads > 0 { iothread_service_main_with_timeout(Duration::from_millis(1000)); } } /// `Debounce` is a simple class which executes one function on a background thread while enqueing /// at most one more. Subsequent execution requests overwrite the enqueued one. It takes an optional /// timeout; if a handler does not finish within the timeout then a new thread is spawned to service /// the remaining request. /// /// Debounce implementation note: we would like to enqueue at most one request, except if a thread /// hangs (e.g. on fs access) then we do not want to block indefinitely - such threads are called /// "abandoned". This is implemented via a monotone uint64 counter, called a token. Every time we /// spawn a thread, we increment the token. When the thread has completed running a work item, it /// compares its token to the active token; if they differ then this thread was abandoned. #[derive(Clone)] pub struct Debounce { timeout: Duration, /// The data shared between [`Debounce`] instances. data: Arc>, } /// The data shared between [`Debounce`] instances. struct DebounceData { /// The (one or none) next enqueued request, overwritten each time a new call to /// [`perform()`](Self::perform) is made. next_req: Option, /// The non-zero token of the current non-abandoned thread or `None` if no thread is running. active_token: Option, /// The next token to use when spawning a thread. next_token: NonZeroU64, /// The start time of the most recently spawned thread or request (if any). start_time: Instant, } fn new_debounce_ffi(timeout_ms: u64) -> Box { Box::new(Debounce::new(Duration::from_millis(timeout_ms))) } impl Debounce { pub fn new(timeout: Duration) -> Self { Self { timeout, data: Arc::new(Mutex::new(DebounceData { next_req: None, active_token: None, next_token: NonZeroU64::new(1).unwrap(), start_time: Instant::now(), })), } } /// Run an iteration in the background with the given thread token. Returns `true` if we handled /// a request or `false` if there were no requests to handle (in which case the debounce thread /// exits). /// /// Note that this method is called from a background thread. fn run_next(&self, token: NonZeroU64) -> bool { let request = { let mut data = self.data.lock().expect("Mutex poisoned!"); if let Some(req) = data.next_req.take() { data.start_time = Instant::now(); req } else { // There is no pending request. Mark this token as no longer running. if Some(token) == data.active_token { data.active_token = None; } return false; } }; // Execute request after unlocking the mutex. (request)(); return true; } /// Enqueue `handler` to be performed on a background thread. If another function is already /// enqueued, this overwrites it and that function will not be executed. /// /// The result is a token which is only of interest to the test suite. pub fn perform(&self, handler: impl FnOnce() + 'static + Send) -> NonZeroU64 { let h = Box::new(handler); self.perform_inner(h) } fn perform_ffi(&self, callback: &cxx::SharedPtr) -> u64 { let callback = callback.clone(); self.perform(move || { callback.invoke(); }) .into() } fn perform_with_completion_ffi( &self, callback: &cxx::SharedPtr, completion: &cxx::SharedPtr, ) -> u64 { let callback = callback.clone(); let completion = completion.clone(); self.perform_with_completion( move || -> crate::ffi::void_ptr { callback.invoke().into() }, move |result| { completion.invoke_with_param(result.into()); }, ) .into() } /// Enqueue `handler` to be performed on a background thread with [`Completion`] `completion` /// to be performed on the main thread. If a function is already enqueued, this overwrites it /// and that function will not be executed. /// /// If the function executes within the optional timeout then `completion` will be invoked on /// the main thread with the result of the evaluated `handler`. /// /// The result is a token which is only of interest to the test suite. pub fn perform_with_completion(&self, handler: H, completion: C) -> NonZeroU64 where H: FnOnce() -> R + 'static + Send, C: FnOnce(R) + 'static + Send, R: 'static + Send, { let h = Box::new(move || { let result = handler(); let c = Box::new(move || { (completion)(result); }); Self::enqueue_main_thread_result(c); }); self.perform_inner(h) } fn perform_inner(&self, handler: WorkItem) -> NonZeroU64 { let mut spawn = false; let active_token = { let mut data = self.data.lock().expect("Mutex poisoned!"); data.next_req = Some(handler); // If we have a timeout and our running thread has exceeded it, abandon that thread. if data.active_token.is_some() && !self.timeout.is_zero() && (Instant::now() - data.start_time > self.timeout) { // Abandon this thread by dissociating its token from this [`Debounce`] instance. data.active_token = None; } if data.active_token.is_none() { // We need to spawn a new thread. Mark the current time so that a new request won't // immediately abandon us and start a new thread too. spawn = true; data.active_token = Some(data.next_token); data.next_token = data.next_token.checked_add(1).unwrap(); data.start_time = Instant::now(); } data.active_token.expect("Something should be active now.") }; // Spawn after unlocking the mutex above. if spawn { // We need to clone the Arc to get it to last for the duration of the 'static lifetime. let debounce = self.clone(); iothread_perform(move || { while debounce.run_next(active_token) { // Keep thread alive/busy. } }); } active_token } /// Static helper to add a [`WorkItem`] to [`MAIN_THREAD_ID`] and signal [`NOTIFY_SIGNALLER`]. fn enqueue_main_thread_result(f: WorkItem) { MAIN_THREAD_QUEUE.lock().expect("Mutex poisoned!").push(f); NOTIFY_SIGNALLER.post(); } } #[test] /// Verify that spawing a thread normally via [`std::thread::spawn()`] causes the calling thread's /// sigmask to be inherited by the newly spawned thread. fn std_thread_inherits_sigmask() { // First change our own thread mask let (saved_set, t1_set) = unsafe { let mut new_set: libc::sigset_t = std::mem::zeroed(); let new_set = &mut new_set as *mut _; libc::sigemptyset(new_set); libc::sigaddset(new_set, libc::SIGILL); // mask bad jump let mut saved_set: libc::sigset_t = std::mem::zeroed(); let result = libc::pthread_sigmask(libc::SIG_BLOCK, new_set, &mut saved_set as *mut _); assert_eq!(result, 0, "Failed to set thread mask!"); // Now get the current set that includes the masked SIGILL let mut t1_set: libc::sigset_t = std::mem::zeroed(); let mut empty_set = std::mem::zeroed(); let empty_set = &mut empty_set as *mut _; libc::sigemptyset(empty_set); let result = libc::pthread_sigmask(libc::SIG_UNBLOCK, empty_set, &mut t1_set as *mut _); assert_eq!(result, 0, "Failed to get own altered thread mask!"); (saved_set, t1_set) }; // Launch a new thread that can access existing variables let t2_set = std::thread::scope(|_| { unsafe { // Set a new thread sigmask and verify that the old one is what we expect it to be let mut new_set: libc::sigset_t = std::mem::zeroed(); let new_set = &mut new_set as *mut _; libc::sigemptyset(new_set); let mut saved_set2: libc::sigset_t = std::mem::zeroed(); let result = libc::pthread_sigmask(libc::SIG_BLOCK, new_set, &mut saved_set2 as *mut _); assert_eq!(result, 0, "Failed to get existing sigmask for new thread"); saved_set2 } }); // Compare the sigset_t values unsafe { let t1_sigset_slice = std::slice::from_raw_parts( &t1_set as *const _ as *const u8, core::mem::size_of::(), ); let t2_sigset_slice = std::slice::from_raw_parts( &t2_set as *const _ as *const u8, core::mem::size_of::(), ); assert_eq!(t1_sigset_slice, t2_sigset_slice); }; // Restore the thread sigset so we don't affect `cargo test`'s multithreaded test harnesses unsafe { let result = libc::pthread_sigmask( libc::SIG_SETMASK, &saved_set as *const _, core::ptr::null_mut(), ); assert_eq!(result, 0, "Failed to restore sigmask!"); } }