| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113 |
- /*
- Copyright (c) 2005-2020 Intel Corporation
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
- */
- #ifndef __TBB_queuing_mutex_H
- #define __TBB_queuing_mutex_H
- #define __TBB_queuing_mutex_H_include_area
- #include "internal/_warning_suppress_enable_notice.h"
- #include <cstring>
- #include "atomic.h"
- #include "tbb_profiling.h"
- namespace tbb {
- //! Queuing mutex with local-only spinning.
- /** @ingroup synchronization */
- class queuing_mutex : internal::mutex_copy_deprecated_and_disabled {
- public:
- //! Construct unacquired mutex.
- queuing_mutex() {
- q_tail = NULL;
- #if TBB_USE_THREADING_TOOLS
- internal_construct();
- #endif
- }
- //! The scoped locking pattern
- /** It helps to avoid the common problem of forgetting to release lock.
- It also nicely provides the "node" for queuing locks. */
- class scoped_lock: internal::no_copy {
- //! Initialize fields to mean "no lock held".
- void initialize() {
- mutex = NULL;
- going = 0;
- #if TBB_USE_ASSERT
- internal::poison_pointer(next);
- #endif /* TBB_USE_ASSERT */
- }
- public:
- //! Construct lock that has not acquired a mutex.
- /** Equivalent to zero-initialization of *this. */
- scoped_lock() {initialize();}
- //! Acquire lock on given mutex.
- scoped_lock( queuing_mutex& m ) {
- initialize();
- acquire(m);
- }
- //! Release lock (if lock is held).
- ~scoped_lock() {
- if( mutex ) release();
- }
- //! Acquire lock on given mutex.
- void __TBB_EXPORTED_METHOD acquire( queuing_mutex& m );
- //! Acquire lock on given mutex if free (i.e. non-blocking)
- bool __TBB_EXPORTED_METHOD try_acquire( queuing_mutex& m );
- //! Release lock.
- void __TBB_EXPORTED_METHOD release();
- private:
- //! The pointer to the mutex owned, or NULL if not holding a mutex.
- queuing_mutex* mutex;
- //! The pointer to the next competitor for a mutex
- scoped_lock *next;
- //! The local spin-wait variable
- /** Inverted (0 - blocked, 1 - acquired the mutex) for the sake of
- zero-initialization. Defining it as an entire word instead of
- a byte seems to help performance slightly. */
- uintptr_t going;
- };
- void __TBB_EXPORTED_METHOD internal_construct();
- // Mutex traits
- static const bool is_rw_mutex = false;
- static const bool is_recursive_mutex = false;
- static const bool is_fair_mutex = true;
- private:
- //! The last competitor requesting the lock
- atomic<scoped_lock*> q_tail;
- };
- __TBB_DEFINE_PROFILING_SET_NAME(queuing_mutex)
- } // namespace tbb
- #include "internal/_warning_suppress_disable_notice.h"
- #undef __TBB_queuing_mutex_H_include_area
- #endif /* __TBB_queuing_mutex_H */
|