/* * Copyright (c) 2018-2021, Andreas Kling * * SPDX-License-Identifier: BSD-2-Clause */ #include #include #include #include namespace Kernel { static SpinLock g_global_futex_lock; static AK::Singleton> g_global_futex_queues; FutexQueue::FutexQueue(FlatPtr user_address_or_offset, Memory::VMObject* vmobject) : m_user_address_or_offset(user_address_or_offset) , m_is_global(vmobject != nullptr) { dbgln_if(FUTEX_DEBUG, "Futex @ {}{}", this, m_is_global ? " (global)" : " (local)"); if (m_is_global) { // Only register for global futexes m_vmobject = vmobject->make_weak_ptr(); vmobject->register_on_deleted_handler(*this); } } FutexQueue::~FutexQueue() { if (m_is_global) { if (auto vmobject = m_vmobject.strong_ref()) vmobject->unregister_on_deleted_handler(*this); } dbgln_if(FUTEX_DEBUG, "~Futex @ {}{}", this, m_is_global ? " (global)" : " (local)"); } void FutexQueue::vmobject_deleted(Memory::VMObject& vmobject) { VERIFY(m_is_global); // If we got called we must be a global futex // Because we're taking ourselves out of the global queue, we need // to make sure we have at last a reference until we're done NonnullRefPtr own_ref(*this); dbgln_if(FUTEX_DEBUG, "Futex::vmobject_deleted @ {}{}", this, m_is_global ? " (global)" : " (local)"); // Because this is called from the VMObject's destructor, getting a // strong_ref in this function is unsafe! m_vmobject = nullptr; // Just to be safe... { ScopedSpinLock lock(g_global_futex_lock); g_global_futex_queues->remove(&vmobject); } bool did_wake_all; auto wake_count = wake_all(did_wake_all); if constexpr (FUTEX_DEBUG) { if (wake_count > 0) dbgln("Futex @ {} unblocked {} waiters due to vmobject free", this, wake_count); } VERIFY(did_wake_all); // No one should be left behind... } void Process::clear_futex_queues_on_exec() { ScopedSpinLock lock(m_futex_lock); for (auto& it : m_futex_queues) { bool did_wake_all; it.value->wake_all(did_wake_all); VERIFY(did_wake_all); // No one should be left behind... } m_futex_queues.clear(); } KResultOr Process::sys$futex(Userspace user_params) { VERIFY_PROCESS_BIG_LOCK_ACQUIRED(this); Syscall::SC_futex_params params; if (!copy_from_user(¶ms, user_params)) return EFAULT; Thread::BlockTimeout timeout; u32 cmd = params.futex_op & FUTEX_CMD_MASK; bool use_realtime_clock = (params.futex_op & FUTEX_CLOCK_REALTIME) != 0; if (use_realtime_clock && cmd != FUTEX_WAIT && cmd != FUTEX_WAIT_BITSET) { return ENOSYS; } switch (cmd) { case FUTEX_WAIT: case FUTEX_WAIT_BITSET: case FUTEX_REQUEUE: case FUTEX_CMP_REQUEUE: { if (params.timeout) { auto timeout_time = copy_time_from_user(params.timeout); if (!timeout_time.has_value()) return EFAULT; bool is_absolute = cmd != FUTEX_WAIT; clockid_t clock_id = use_realtime_clock ? CLOCK_REALTIME_COARSE : CLOCK_MONOTONIC_COARSE; timeout = Thread::BlockTimeout(is_absolute, &timeout_time.value(), nullptr, clock_id); } if (cmd == FUTEX_WAIT_BITSET && params.val3 == FUTEX_BITSET_MATCH_ANY) cmd = FUTEX_WAIT; break; case FUTEX_WAKE_BITSET: if (params.val3 == FUTEX_BITSET_MATCH_ANY) cmd = FUTEX_WAKE; break; } } bool is_private = (params.futex_op & FUTEX_PRIVATE_FLAG) != 0; auto& queue_lock = is_private ? m_futex_lock : g_global_futex_lock; auto user_address_or_offset = FlatPtr(params.userspace_address); auto user_address_or_offset2 = FlatPtr(params.userspace_address2); // If this is a global lock, look up the underlying VMObject *before* // acquiring the queue lock RefPtr vmobject, vmobject2; if (!is_private) { auto region = address_space().find_region_containing(Memory::VirtualRange { VirtualAddress { user_address_or_offset }, sizeof(u32) }); if (!region) return EFAULT; vmobject = region->vmobject(); user_address_or_offset = region->offset_in_vmobject_from_vaddr(VirtualAddress(user_address_or_offset)); switch (cmd) { case FUTEX_REQUEUE: case FUTEX_CMP_REQUEUE: case FUTEX_WAKE_OP: { auto region2 = address_space().find_region_containing(Memory::VirtualRange { VirtualAddress { user_address_or_offset2 }, sizeof(u32) }); if (!region2) return EFAULT; vmobject2 = region2->vmobject(); user_address_or_offset2 = region->offset_in_vmobject_from_vaddr(VirtualAddress(user_address_or_offset2)); break; } } } auto find_global_futex_queues = [&](Memory::VMObject& vmobject, bool create_if_not_found) -> FutexQueues* { auto& global_queues = *g_global_futex_queues; auto it = global_queues.find(&vmobject); if (it != global_queues.end()) return &it->value; if (create_if_not_found) { // TODO: is there a better way than setting and finding it again? auto result = global_queues.set(&vmobject, {}); VERIFY(result == AK::HashSetResult::InsertedNewEntry); it = global_queues.find(&vmobject); VERIFY(it != global_queues.end()); return &it->value; } return nullptr; }; auto find_futex_queue = [&](Memory::VMObject* vmobject, FlatPtr user_address_or_offset, bool create_if_not_found, bool* did_create = nullptr) -> RefPtr { VERIFY(is_private || vmobject); VERIFY(!create_if_not_found || did_create != nullptr); auto* queues = is_private ? &m_futex_queues : find_global_futex_queues(*vmobject, create_if_not_found); if (!queues) return {}; auto it = queues->find(user_address_or_offset); if (it != queues->end()) return it->value; if (create_if_not_found) { *did_create = true; auto futex_queue = adopt_ref(*new FutexQueue(user_address_or_offset, vmobject)); auto result = queues->set(user_address_or_offset, futex_queue); VERIFY(result == AK::HashSetResult::InsertedNewEntry); return futex_queue; } return {}; }; auto remove_futex_queue = [&](Memory::VMObject* vmobject, FlatPtr user_address_or_offset) { auto* queues = is_private ? &m_futex_queues : find_global_futex_queues(*vmobject, false); if (queues) { if (auto it = queues->find(user_address_or_offset); it != queues->end()) { if (it->value->try_remove()) { it->value->did_remove(); queues->remove(it); } } if (!is_private && queues->is_empty()) g_global_futex_queues->remove(vmobject); } }; auto do_wake = [&](Memory::VMObject* vmobject, FlatPtr user_address_or_offset, u32 count, Optional bitmask) -> int { if (count == 0) return 0; ScopedSpinLock lock(queue_lock); auto futex_queue = find_futex_queue(vmobject, user_address_or_offset, false); if (!futex_queue) return 0; bool is_empty; u32 woke_count = futex_queue->wake_n(count, bitmask, is_empty); if (is_empty) { // If there are no more waiters, we want to get rid of the futex! remove_futex_queue(vmobject, user_address_or_offset); } return (int)woke_count; }; auto do_wait = [&](u32 bitset) -> int { bool did_create; RefPtr futex_queue; do { auto user_value = user_atomic_load_relaxed(params.userspace_address); if (!user_value.has_value()) return EFAULT; if (user_value.value() != params.val) { dbgln_if(FUTEX_DEBUG, "futex wait: EAGAIN. user value: {:p} @ {:p} != val: {}", user_value.value(), params.userspace_address, params.val); return EAGAIN; } atomic_thread_fence(AK::MemoryOrder::memory_order_acquire); ScopedSpinLock lock(queue_lock); did_create = false; futex_queue = find_futex_queue(vmobject.ptr(), user_address_or_offset, true, &did_create); VERIFY(futex_queue); // We need to try again if we didn't create this queue and the existing queue // was removed before we were able to queue an imminent wait. } while (!did_create && !futex_queue->queue_imminent_wait()); // We must not hold the lock before blocking. But we have a reference // to the FutexQueue so that we can keep it alive. Thread::BlockResult block_result = futex_queue->wait_on(timeout, bitset); ScopedSpinLock lock(queue_lock); if (futex_queue->is_empty_and_no_imminent_waits()) { // If there are no more waiters, we want to get rid of the futex! remove_futex_queue(vmobject, user_address_or_offset); } if (block_result == Thread::BlockResult::InterruptedByTimeout) { return ETIMEDOUT; } return 0; }; auto do_requeue = [&](Optional val3) -> int { auto user_value = user_atomic_load_relaxed(params.userspace_address); if (!user_value.has_value()) return EFAULT; if (val3.has_value() && val3.value() != user_value.value()) return EAGAIN; atomic_thread_fence(AK::MemoryOrder::memory_order_acquire); int woken_or_requeued = 0; ScopedSpinLock lock(queue_lock); if (auto futex_queue = find_futex_queue(vmobject.ptr(), user_address_or_offset, false)) { RefPtr target_futex_queue; bool is_empty, is_target_empty; woken_or_requeued = futex_queue->wake_n_requeue( params.val, [&]() -> FutexQueue* { // NOTE: futex_queue's lock is being held while this callback is called // The reason we're doing this in a callback is that we don't want to always // create a target queue, only if we actually have anything to move to it! target_futex_queue = find_futex_queue(vmobject2.ptr(), user_address_or_offset2, true); return target_futex_queue.ptr(); }, params.val2, is_empty, is_target_empty); if (is_empty) remove_futex_queue(vmobject, user_address_or_offset); if (is_target_empty && target_futex_queue) remove_futex_queue(vmobject2, user_address_or_offset2); } return woken_or_requeued; }; switch (cmd) { case FUTEX_WAIT: return do_wait(0); case FUTEX_WAKE: return do_wake(vmobject.ptr(), user_address_or_offset, params.val, {}); case FUTEX_WAKE_OP: { Optional oldval; u32 op_arg = _FUTEX_OP_ARG(params.val3); auto op = _FUTEX_OP(params.val3); if (op & FUTEX_OP_ARG_SHIFT) { op_arg = 1 << op_arg; op &= FUTEX_OP_ARG_SHIFT; } atomic_thread_fence(AK::MemoryOrder::memory_order_release); switch (op) { case FUTEX_OP_SET: oldval = user_atomic_exchange_relaxed(params.userspace_address2, op_arg); break; case FUTEX_OP_ADD: oldval = user_atomic_fetch_add_relaxed(params.userspace_address2, op_arg); break; case FUTEX_OP_OR: oldval = user_atomic_fetch_or_relaxed(params.userspace_address2, op_arg); break; case FUTEX_OP_ANDN: oldval = user_atomic_fetch_and_not_relaxed(params.userspace_address2, op_arg); break; case FUTEX_OP_XOR: oldval = user_atomic_fetch_xor_relaxed(params.userspace_address2, op_arg); break; default: return EINVAL; } if (!oldval.has_value()) return EFAULT; atomic_thread_fence(AK::MemoryOrder::memory_order_acquire); int result = do_wake(vmobject.ptr(), user_address_or_offset, params.val, {}); if (params.val2 > 0) { bool compare_result; switch (_FUTEX_CMP(params.val3)) { case FUTEX_OP_CMP_EQ: compare_result = (oldval.value() == _FUTEX_CMP_ARG(params.val3)); break; case FUTEX_OP_CMP_NE: compare_result = (oldval.value() != _FUTEX_CMP_ARG(params.val3)); break; case FUTEX_OP_CMP_LT: compare_result = (oldval.value() < _FUTEX_CMP_ARG(params.val3)); break; case FUTEX_OP_CMP_LE: compare_result = (oldval.value() <= _FUTEX_CMP_ARG(params.val3)); break; case FUTEX_OP_CMP_GT: compare_result = (oldval.value() > _FUTEX_CMP_ARG(params.val3)); break; case FUTEX_OP_CMP_GE: compare_result = (oldval.value() >= _FUTEX_CMP_ARG(params.val3)); break; default: return EINVAL; } if (compare_result) result += do_wake(vmobject2.ptr(), user_address_or_offset2, params.val2, {}); } return result; } case FUTEX_REQUEUE: return do_requeue({}); case FUTEX_CMP_REQUEUE: return do_requeue(params.val3); case FUTEX_WAIT_BITSET: VERIFY(params.val3 != FUTEX_BITSET_MATCH_ANY); // we should have turned it into FUTEX_WAIT if (params.val3 == 0) return EINVAL; return do_wait(params.val3); case FUTEX_WAKE_BITSET: VERIFY(params.val3 != FUTEX_BITSET_MATCH_ANY); // we should have turned it into FUTEX_WAKE if (params.val3 == 0) return EINVAL; return do_wake(vmobject.ptr(), user_address_or_offset, params.val, params.val3); } return ENOSYS; } }