X-Git-Url: https://git.sesse.net/?p=stockfish;a=blobdiff_plain;f=src%2Fthread.cpp;h=5d2b39f57c8c4791dd6b28975448a55d89fa2622;hp=c7cabeb7e131afcdc75d4a036666c8c6407449a7;hb=9a542d96981e6cb45b6b01f17258a078cf27da36;hpb=92d70fb6677f4ed0568aa50f239ceea2d8a1cd97 diff --git a/src/thread.cpp b/src/thread.cpp index c7cabeb7..5d2b39f5 100644 --- a/src/thread.cpp +++ b/src/thread.cpp @@ -132,7 +132,7 @@ void ThreadsManager::init() { // Allocate pawn and material hash tables for main thread init_hash_tables(); - lock_init(&mpLock); + lock_init(&threadsLock); // Initialize thread and split point locks for (int i = 0; i < MAX_THREADS; i++) @@ -193,7 +193,7 @@ void ThreadsManager::exit() { lock_destroy(&(threads[i].splitPoints[j].lock)); } - lock_destroy(&mpLock); + lock_destroy(&threadsLock); } @@ -240,7 +240,7 @@ bool ThreadsManager::available_slave_exists(int master) const { template void ThreadsManager::split(Position& pos, SearchStack* ss, Value* alpha, const Value beta, Value* bestValue, Depth depth, Move threatMove, - int moveCount, MovePicker* mp, bool pvNode) { + int moveCount, MovePicker* mp, int nodeType) { assert(pos.is_ok()); assert(*bestValue >= -VALUE_INFINITE); assert(*bestValue <= *alpha); @@ -253,19 +253,12 @@ void ThreadsManager::split(Position& pos, SearchStack* ss, Value* alpha, const V int i, master = pos.thread(); Thread& masterThread = threads[master]; - lock_grab(&mpLock); - - // If no other thread is available to help us, or if we have too many - // active split points, don't split. - if ( !available_slave_exists(master) - || masterThread.activeSplitPoints >= MAX_ACTIVE_SPLIT_POINTS) - { - lock_release(&mpLock); + // If we already have too many active split points, don't split + if (masterThread.activeSplitPoints >= MAX_ACTIVE_SPLIT_POINTS) return; - } // Pick the next available split point object from the split point stack - SplitPoint& splitPoint = masterThread.splitPoints[masterThread.activeSplitPoints++]; + SplitPoint& splitPoint = masterThread.splitPoints[masterThread.activeSplitPoints]; // Initialize the split point object splitPoint.parent = masterThread.splitPoint; @@ -275,7 +268,7 @@ void ThreadsManager::split(Position& pos, SearchStack* ss, Value* alpha, const V splitPoint.threatMove = threatMove; splitPoint.alpha = *alpha; splitPoint.beta = beta; - splitPoint.pvNode = pvNode; + splitPoint.nodeType = nodeType; splitPoint.bestValue = *bestValue; splitPoint.mp = mp; splitPoint.moveCount = moveCount; @@ -285,27 +278,33 @@ void ThreadsManager::split(Position& pos, SearchStack* ss, Value* alpha, const V for (i = 0; i < activeThreads; i++) splitPoint.is_slave[i] = false; - masterThread.splitPoint = &splitPoint; - // If we are here it means we are not available - assert(masterThread.state != Thread::AVAILABLE); + assert(masterThread.state == Thread::SEARCHING); + + int booked = 0; - int workersCnt = 1; // At least the master is included + // Try to allocate available threads setting state to Thread::BOOKED, this + // must be done under lock protection to avoid concurrent allocation of + // the same slave by another master. + lock_grab(&threadsLock); - // Allocate available threads setting state to THREAD_BOOKED - for (i = 0; !Fake && i < activeThreads && workersCnt < maxThreadsPerSplitPoint; i++) + for (i = 0; !Fake && i < activeThreads && booked < maxThreadsPerSplitPoint; i++) if (i != master && threads[i].is_available_to(master)) { threads[i].state = Thread::BOOKED; threads[i].splitPoint = &splitPoint; splitPoint.is_slave[i] = true; - workersCnt++; + booked++; } - assert(Fake || workersCnt > 1); + lock_release(&threadsLock); + + // We failed to allocate even one slave, return + if (!Fake && !booked) + return; - // We can release the lock because slave threads are already booked and master is not available - lock_release(&mpLock); + masterThread.activeSplitPoints++; + masterThread.splitPoint = &splitPoint; // Tell the threads that they have work to do. This will make them leave // their idle loop. @@ -314,7 +313,8 @@ void ThreadsManager::split(Position& pos, SearchStack* ss, Value* alpha, const V { assert(i == master || threads[i].state == Thread::BOOKED); - threads[i].state = Thread::WORKISWAITING; // This makes the slave to exit from idle_loop() + // This makes the slave to exit from idle_loop() + threads[i].state = Thread::WORKISWAITING; if (useSleepingThreads && i != master) threads[i].wake_up(); @@ -328,18 +328,22 @@ void ThreadsManager::split(Position& pos, SearchStack* ss, Value* alpha, const V idle_loop(master, &splitPoint); // We have returned from the idle loop, which means that all threads are - // finished. Update alpha and bestValue, and return. - lock_grab(&mpLock); + // finished. Update alpha and bestValue, and return. Note that changing + // state and decreasing activeSplitPoints is done under lock protection + // to avoid a race with Thread::is_available_to(). + lock_grab(&threadsLock); - *alpha = splitPoint.alpha; - *bestValue = splitPoint.bestValue; + masterThread.state = Thread::SEARCHING; masterThread.activeSplitPoints--; masterThread.splitPoint = splitPoint.parent; - pos.set_nodes_searched(pos.nodes_searched() + splitPoint.nodes); - lock_release(&mpLock); + lock_release(&threadsLock); + + *alpha = splitPoint.alpha; + *bestValue = splitPoint.bestValue; + pos.set_nodes_searched(pos.nodes_searched() + splitPoint.nodes); } // Explicit template instantiations -template void ThreadsManager::split(Position&, SearchStack*, Value*, const Value, Value*, Depth, Move, int, MovePicker*, bool); -template void ThreadsManager::split(Position&, SearchStack*, Value*, const Value, Value*, Depth, Move, int, MovePicker*, bool); +template void ThreadsManager::split(Position&, SearchStack*, Value*, const Value, Value*, Depth, Move, int, MovePicker*, int); +template void ThreadsManager::split(Position&, SearchStack*, Value*, const Value, Value*, Depth, Move, int, MovePicker*, int);