mirror of
https://github.com/sockspls/badfish
synced 2025-04-30 00:33:09 +00:00
Improve spinlock implementation
Calling lock.test_and_set() in a tight loop creates expensive memory synchronizations among processors and penalize other running threads. So syncronize only only once at the beginning with fetch_sub() and then loop on a simple load() that puts much less pressure on the system. Reported about 2-3% speed up on various systems. Patch by Ronald de Man. No functional change.
This commit is contained in:
parent
38112060dc
commit
d3d26a94b3
1 changed files with 9 additions and 5 deletions
14
src/thread.h
14
src/thread.h
|
@ -39,16 +39,20 @@ const size_t MAX_THREADS = 128;
|
|||
const size_t MAX_SPLITPOINTS_PER_THREAD = 8;
|
||||
const size_t MAX_SLAVES_PER_SPLITPOINT = 4;
|
||||
|
||||
/// Spinlock class wraps low level atomic operations to provide spin lock functionality
|
||||
|
||||
/// Spinlock class wraps low level atomic operations to provide a spin lock
|
||||
|
||||
class Spinlock {
|
||||
|
||||
std::atomic_flag lock;
|
||||
std::atomic_int lock;
|
||||
|
||||
public:
|
||||
Spinlock() { std::atomic_flag_clear(&lock); }
|
||||
void acquire() { while (lock.test_and_set(std::memory_order_acquire)) {} }
|
||||
void release() { lock.clear(std::memory_order_release); }
|
||||
Spinlock() { lock = 1; } // Init here to workaround a bug with MSVC 2013
|
||||
void acquire() {
|
||||
while (lock.fetch_sub(1, std::memory_order_acquire) != 1)
|
||||
while (lock.load(std::memory_order_relaxed) <= 0) {}
|
||||
}
|
||||
void release() { lock.store(1, std::memory_order_release); }
|
||||
};
|
||||
|
||||
|
||||
|
|
Loading…
Add table
Reference in a new issue