| // Copyright 2016 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "base/task_scheduler/scheduler_lock_impl.h" |
| |
| #include <algorithm> |
| #include <unordered_map> |
| #include <vector> |
| |
| #include "base/lazy_instance.h" |
| #include "base/logging.h" |
| #include "base/synchronization/condition_variable.h" |
| #include "base/threading/platform_thread.h" |
| #include "base/threading/thread_local_storage.h" |
| |
| namespace base { |
| namespace internal { |
| |
| namespace { |
| |
| class SafeAcquisitionTracker { |
| public: |
| SafeAcquisitionTracker() : tls_acquired_locks_(&OnTLSDestroy) {} |
| |
| void RegisterLock( |
| const SchedulerLockImpl* const lock, |
| const SchedulerLockImpl* const predecessor) { |
| DCHECK_NE(lock, predecessor) << "Reentrant locks are unsupported."; |
| AutoLock auto_lock(allowed_predecessor_map_lock_); |
| allowed_predecessor_map_[lock] = predecessor; |
| AssertSafePredecessor(lock); |
| } |
| |
| void UnregisterLock(const SchedulerLockImpl* const lock) { |
| AutoLock auto_lock(allowed_predecessor_map_lock_); |
| allowed_predecessor_map_.erase(lock); |
| } |
| |
| void RecordAcquisition(const SchedulerLockImpl* const lock) { |
| AssertSafeAcquire(lock); |
| GetAcquiredLocksOnCurrentThread()->push_back(lock); |
| } |
| |
| void RecordRelease(const SchedulerLockImpl* const lock) { |
| LockVector* acquired_locks = GetAcquiredLocksOnCurrentThread(); |
| const auto iter_at_lock = |
| std::find(acquired_locks->begin(), acquired_locks->end(), lock); |
| DCHECK(iter_at_lock != acquired_locks->end()); |
| acquired_locks->erase(iter_at_lock); |
| } |
| |
| private: |
| using LockVector = std::vector<const SchedulerLockImpl*>; |
| using PredecessorMap = std::unordered_map< |
| const SchedulerLockImpl*, const SchedulerLockImpl*>; |
| |
| // This asserts that the lock is safe to acquire. This means that this should |
| // be run before actually recording the acquisition. |
| void AssertSafeAcquire(const SchedulerLockImpl* const lock) { |
| const LockVector* acquired_locks = GetAcquiredLocksOnCurrentThread(); |
| |
| // If the thread currently holds no locks, this is inherently safe. |
| if (acquired_locks->empty()) |
| return; |
| |
| // Otherwise, make sure that the previous lock acquired is an allowed |
| // predecessor. |
| AutoLock auto_lock(allowed_predecessor_map_lock_); |
| // Using at() is exception-safe here as |lock| was registered already. |
| const SchedulerLockImpl* allowed_predecessor = |
| allowed_predecessor_map_.at(lock); |
| DCHECK_EQ(acquired_locks->back(), allowed_predecessor); |
| } |
| |
| // Asserts that |lock|'s registered predecessor is safe. Because |
| // SchedulerLocks are registered at construction time and any predecessor |
| // specified on a SchedulerLock must already exist, the first registered |
| // SchedulerLock in a potential chain must have a null predecessor and is thus |
| // cycle-free. Any subsequent SchedulerLock with a predecessor must come from |
| // the set of registered SchedulerLocks. Since the registered SchedulerLocks |
| // only contain cycle-free SchedulerLocks, this subsequent SchedulerLock is |
| // itself cycle-free and may be safely added to the registered SchedulerLock |
| // set. |
| void AssertSafePredecessor(const SchedulerLockImpl* lock) const { |
| allowed_predecessor_map_lock_.AssertAcquired(); |
| // Using at() is exception-safe here as |lock| was registered already. |
| const SchedulerLockImpl* predecessor = allowed_predecessor_map_.at(lock); |
| if (predecessor) { |
| DCHECK(allowed_predecessor_map_.find(predecessor) != |
| allowed_predecessor_map_.end()) |
| << "SchedulerLock was registered before its predecessor. " |
| << "Potential cycle detected"; |
| } |
| } |
| |
| LockVector* GetAcquiredLocksOnCurrentThread() { |
| if (!tls_acquired_locks_.Get()) |
| tls_acquired_locks_.Set(new LockVector); |
| |
| return reinterpret_cast<LockVector*>(tls_acquired_locks_.Get()); |
| } |
| |
| static void OnTLSDestroy(void* value) { |
| delete reinterpret_cast<LockVector*>(value); |
| } |
| |
| // Synchronizes access to |allowed_predecessor_map_|. |
| Lock allowed_predecessor_map_lock_; |
| |
| // A map of allowed predecessors. |
| PredecessorMap allowed_predecessor_map_; |
| |
| // A thread-local slot holding a vector of locks currently acquired on the |
| // current thread. |
| ThreadLocalStorage::Slot tls_acquired_locks_; |
| |
| DISALLOW_COPY_AND_ASSIGN(SafeAcquisitionTracker); |
| }; |
| |
| LazyInstance<SafeAcquisitionTracker>::Leaky g_safe_acquisition_tracker = |
| LAZY_INSTANCE_INITIALIZER; |
| |
| } // namespace |
| |
| SchedulerLockImpl::SchedulerLockImpl() : SchedulerLockImpl(nullptr) {} |
| |
| SchedulerLockImpl::SchedulerLockImpl(const SchedulerLockImpl* predecessor) { |
| g_safe_acquisition_tracker.Get().RegisterLock(this, predecessor); |
| } |
| |
| SchedulerLockImpl::~SchedulerLockImpl() { |
| g_safe_acquisition_tracker.Get().UnregisterLock(this); |
| } |
| |
| void SchedulerLockImpl::Acquire() { |
| lock_.Acquire(); |
| g_safe_acquisition_tracker.Get().RecordAcquisition(this); |
| } |
| |
| void SchedulerLockImpl::Release() { |
| lock_.Release(); |
| g_safe_acquisition_tracker.Get().RecordRelease(this); |
| } |
| |
| void SchedulerLockImpl::AssertAcquired() const { |
| lock_.AssertAcquired(); |
| } |
| |
| std::unique_ptr<ConditionVariable> |
| SchedulerLockImpl::CreateConditionVariable() { |
| return std::unique_ptr<ConditionVariable>(new ConditionVariable(&lock_)); |
| } |
| |
| } // namespace internal |
| } // base |