1
0
Fork 0
mirror of https://github.com/sockspls/badfish synced 2025-07-11 19:49:14 +00:00

Introduce yielding spin locks

Idea and original implementation by Stephane Nicolet

7 threads 15+0.05
ELO: 3.54 +-2.9 (95%) LOS: 99.2%
Total: 17971 W: 2976 L: 2793 D: 12202

There is no functional change in single thread mode
This commit is contained in:
Joona Kiiski 2015-03-12 20:36:32 +00:00
parent 558b0c848c
commit d71f707040
3 changed files with 21 additions and 6 deletions

View file

@ -1696,7 +1696,7 @@ void Thread::idle_loop() {
if ( sp->allSlavesSearching if ( sp->allSlavesSearching
&& sp->slavesMask.count() < MAX_SLAVES_PER_SPLITPOINT) && sp->slavesMask.count() < MAX_SLAVES_PER_SPLITPOINT)
{ {
mutex.lock(); allocMutex.lock();
if (can_join(sp)) if (can_join(sp))
{ {
@ -1705,7 +1705,7 @@ void Thread::idle_loop() {
searching = true; searching = true;
} }
mutex.unlock(); allocMutex.unlock();
} }
sp->mutex.unlock(); sp->mutex.unlock();

View file

@ -174,17 +174,18 @@ void Thread::split(Position& pos, Stack* ss, Value alpha, Value beta, Value* bes
while ( sp.slavesMask.count() < MAX_SLAVES_PER_SPLITPOINT while ( sp.slavesMask.count() < MAX_SLAVES_PER_SPLITPOINT
&& (slave = Threads.available_slave(&sp)) != nullptr) && (slave = Threads.available_slave(&sp)) != nullptr)
{ {
slave->mutex.lock(); slave->allocMutex.lock();
if (slave->can_join(activeSplitPoint)) if (slave->can_join(activeSplitPoint))
{ {
activeSplitPoint->slavesMask.set(slave->idx); activeSplitPoint->slavesMask.set(slave->idx);
slave->activeSplitPoint = activeSplitPoint; slave->activeSplitPoint = activeSplitPoint;
slave->searching = true; slave->searching = true;
slave->sleepCondition.notify_one(); // Could be sleeping
} }
slave->mutex.unlock(); slave->allocMutex.unlock();
slave->notify_one(); // Could be sleeping
} }
// Everything is set up. The master thread enters the idle loop, from which // Everything is set up. The master thread enters the idle loop, from which

View file

@ -40,6 +40,19 @@ const size_t MAX_THREADS = 128;
const size_t MAX_SPLITPOINTS_PER_THREAD = 8; const size_t MAX_SPLITPOINTS_PER_THREAD = 8;
const size_t MAX_SLAVES_PER_SPLITPOINT = 4; const size_t MAX_SLAVES_PER_SPLITPOINT = 4;
class Spinlock {
std::atomic_int _lock;
public:
Spinlock() { _lock = 1; } // Init here to workaround a bug with MSVC 2013
void lock() {
while (_lock.fetch_sub(1, std::memory_order_acquire) != 1)
for (int cnt = 0; _lock.load(std::memory_order_relaxed) <= 0; ++cnt)
if (cnt >= 10000) std::this_thread::yield(); // Be nice to hyperthreading
}
void unlock() { _lock.store(1, std::memory_order_release); }
};
/// SplitPoint struct stores information shared by the threads searching in /// SplitPoint struct stores information shared by the threads searching in
/// parallel below the same split point. It is populated at splitting time. /// parallel below the same split point. It is populated at splitting time.
@ -60,7 +73,7 @@ struct SplitPoint {
SplitPoint* parentSplitPoint; SplitPoint* parentSplitPoint;
// Shared variable data // Shared variable data
Mutex mutex; Spinlock mutex;
std::bitset<MAX_THREADS> slavesMask; std::bitset<MAX_THREADS> slavesMask;
volatile bool allSlavesSearching; volatile bool allSlavesSearching;
volatile uint64_t nodes; volatile uint64_t nodes;
@ -114,6 +127,7 @@ struct Thread : public ThreadBase {
SplitPoint* volatile activeSplitPoint; SplitPoint* volatile activeSplitPoint;
volatile size_t splitPointsSize; volatile size_t splitPointsSize;
volatile bool searching; volatile bool searching;
Spinlock allocMutex;
}; };