// Futility margin
Value futility_margin(Depth d, bool improving) {
- return Value(234 * (d - improving));
+ return Value(214 * (d - improving));
}
// Reductions lookup table, initialized at startup
Depth reduction(bool i, Depth d, int mn) {
int r = Reductions[d] * Reductions[mn];
- return (r + 503) / 1024 + (!i && r > 915);
+ return (r + 534) / 1024 + (!i && r > 904);
}
constexpr int futility_move_count(bool improving, Depth depth) {
// History and stats update bonus, based on depth
int stat_bonus(Depth d) {
- return d > 14 ? 66 : 6 * d * d + 231 * d - 206;
+ return d > 14 ? 73 : 6 * d * d + 229 * d - 215;
}
// Add a small random component to draw evaluations to avoid 3-fold blindness
Move best = MOVE_NONE;
};
- // Breadcrumbs are used to mark nodes as being searched by a given thread
- struct Breadcrumb {
- std::atomic<Thread*> thread;
- std::atomic<Key> key;
- };
- std::array<Breadcrumb, 1024> breadcrumbs;
-
- // ThreadHolding structure keeps track of which thread left breadcrumbs at the given
- // node for potential reductions. A free node will be marked upon entering the moves
- // loop by the constructor, and unmarked upon leaving that loop by the destructor.
- struct ThreadHolding {
- explicit ThreadHolding(Thread* thisThread, Key posKey, int ply) {
- location = ply < 8 ? &breadcrumbs[posKey & (breadcrumbs.size() - 1)] : nullptr;
- otherThread = false;
- owning = false;
- if (location)
- {
- // See if another already marked this location, if not, mark it ourselves
- Thread* tmp = (*location).thread.load(std::memory_order_relaxed);
- if (tmp == nullptr)
- {
- (*location).thread.store(thisThread, std::memory_order_relaxed);
- (*location).key.store(posKey, std::memory_order_relaxed);
- owning = true;
- }
- else if ( tmp != thisThread
- && (*location).key.load(std::memory_order_relaxed) == posKey)
- otherThread = true;
- }
- }
-
- ~ThreadHolding() {
- if (owning) // Free the marked location
- (*location).thread.store(nullptr, std::memory_order_relaxed);
- }
-
- bool marked() { return otherThread; }
-
- private:
- Breadcrumb* location;
- bool otherThread, owning;
- };
-
template <NodeType NT>
Value search(Position& pos, Stack* ss, Value alpha, Value beta, Depth depth, bool cutNode);
uint64_t perft(Position& pos, Depth depth) {
StateInfo st;
- ASSERT_ALIGNED(&st, Eval::NNUE::kCacheLineSize);
+ ASSERT_ALIGNED(&st, Eval::NNUE::CacheLineSize);
uint64_t cnt, nodes = 0;
const bool leaf = (depth == 2);
void Search::init() {
for (int i = 1; i < MAX_MOVES; ++i)
- Reductions[i] = int((21.3 + 2 * std::log(Threads.size())) * std::log(i + 0.25 * std::log(i)));
+ Reductions[i] = int(21.3 * std::log(i + 0.25 * std::log(i)));
}
// Start with a small aspiration window and, in the case of a fail
// high/low, re-search with a bigger window until we don't fail
// high/low anymore.
- failedHighCnt = 0;
+ int failedHighCnt = 0;
while (true)
{
Depth adjustedDepth = std::max(1, rootDepth - failedHighCnt - searchAgainCounter);
Move pv[MAX_PLY+1], capturesSearched[32], quietsSearched[64];
StateInfo st;
- ASSERT_ALIGNED(&st, Eval::NNUE::kCacheLineSize);
+ ASSERT_ALIGNED(&st, Eval::NNUE::CacheLineSize);
TTEntry* tte;
Key posKey;
if ((ss-1)->currentMove != MOVE_NULL)
ss->staticEval = eval = evaluate(pos);
else
- ss->staticEval = eval = -(ss-1)->staticEval + 2 * Tempo;
+ ss->staticEval = eval = -(ss-1)->staticEval;
// Save static evaluation into transposition table
tte->save(posKey, VALUE_NONE, ss->ttPv, BOUND_NONE, DEPTH_NONE, MOVE_NONE, eval);
// Use static evaluation difference to improve quiet move ordering
if (is_ok((ss-1)->currentMove) && !(ss-1)->inCheck && !priorCapture)
{
- int bonus = std::clamp(-depth * 4 * int((ss-1)->staticEval + ss->staticEval - 2 * Tempo), -1000, 1000);
+ int bonus = std::clamp(-depth * 4 * int((ss-1)->staticEval + ss->staticEval), -1000, 1000);
thisThread->mainHistory[~us][from_to((ss-1)->currentMove)] << bonus;
}
// Step 8. Null move search with verification search (~40 Elo)
if ( !PvNode
&& (ss-1)->currentMove != MOVE_NULL
- && (ss-1)->statScore < 24185
+ && (ss-1)->statScore < 23767
&& eval >= beta
&& eval >= ss->staticEval
- && ss->staticEval >= beta - 24 * depth - 34 * improving + 162 * ss->ttPv + 159
+ && ss->staticEval >= beta - 20 * depth - 22 * improving + 168 * ss->ttPv + 159
&& !excludedMove
&& pos.non_pawn_material(us)
&& (ss->ply >= thisThread->nmpMinPly || us != thisThread->nmpColor))
assert(eval - beta >= 0);
// Null move dynamic reduction based on depth and value
- Depth R = (1062 + 68 * depth) / 256 + std::min(int(eval - beta) / 190, 3);
+ Depth R = (1090 + 81 * depth) / 256 + std::min(int(eval - beta) / 205, 3);
ss->currentMove = MOVE_NULL;
ss->continuationHistory = &thisThread->continuationHistory[0][0][NO_PIECE][0];
probCutBeta = beta + 209 - 44 * improving;
- // Step 9. ProbCut (~10 Elo)
+ // Step 9. ProbCut (~4 Elo)
// If we have a good enough capture and a reduced search returns a value
// much above beta, we can (almost) safely prune the previous move.
if ( !PvNode
&& ttValue != VALUE_NONE
&& ttValue < probCutBeta))
{
- // if ttMove is a capture and value from transposition table is good enough produce probCut
- // cutoff without digging into actual probCut search
- if ( ss->ttHit
- && tte->depth() >= depth - 3
- && ttValue != VALUE_NONE
- && ttValue >= probCutBeta
- && ttMove
- && pos.capture_or_promotion(ttMove))
- return probCutBeta;
-
assert(probCutBeta < VALUE_INFINITE);
MovePicker mp(pos, ttMove, probCutBeta - ss->staticEval, &captureHistory);
ttCapture = ttMove && pos.capture_or_promotion(ttMove);
// Step 11. A small Probcut idea, when we are in check
- probCutBeta = beta + 400;
+ probCutBeta = beta + 409;
if ( ss->inCheck
&& !PvNode
&& depth >= 4
value = bestValue;
singularQuietLMR = moveCountPruning = false;
- // Mark this node as being searched
- ThreadHolding th(thisThread, posKey, ss->ply);
+ // Indicate PvNodes that will probably fail low if the node was searched
+ // at a depth equal or greater than the current depth, and the result of this search was a fail low.
+ bool likelyFailLow = PvNode
+ && ttMove
+ && (tte->bound() & BOUND_UPPER)
+ && tte->depth() >= depth;
// Step 12. Loop through all pseudo-legal moves until no moves remain
// or a beta cutoff occurs.
movedPiece = pos.moved_piece(move);
givesCheck = pos.gives_check(move);
- // Indicate PvNodes that will probably fail low if node was searched with non-PV search
- // at depth equal or greater to current depth and result of this search was far below alpha
- bool likelyFailLow = PvNode
- && ttMove
- && (tte->bound() & BOUND_UPPER)
- && ttValue < alpha + 200 + 100 * depth
- && tte->depth() >= depth;
-
// Calculate new depth for this move
newDepth = depth - 1;
}
else
{
- // Countermoves based pruning (~20 Elo)
- if ( lmrDepth < 4 + ((ss-1)->statScore > 0 || (ss-1)->moveCount == 1)
+ // Continuation history based pruning (~20 Elo)
+ if ( lmrDepth < 5
&& (*contHist[0])[movedPiece][to_sq(move)] < CounterMovePruneThreshold
&& (*contHist[1])[movedPiece][to_sq(move)] < CounterMovePruneThreshold)
continue;
{
extension = 1;
singularQuietLMR = !ttCapture;
+ if (!PvNode && value < singularBeta - 93)
+ extension = 2;
}
// Multi-cut pruning
}
}
- // Check extension (~2 Elo)
- else if ( givesCheck
- && (pos.is_discovered_check_on_king(~us, move) || pos.see_ge(move)))
- extension = 1;
-
- // Last captures extension
- else if ( PieceValue[EG][pos.captured_piece()] > PawnValueEg
- && pos.non_pawn_material() <= 2 * RookValueMg)
- extension = 1;
-
// Add extension to new depth
newDepth += extension;
if ( depth >= 3
&& moveCount > 1 + 2 * rootNode
&& ( !captureOrPromotion
- || moveCountPruning
- || ss->staticEval + PieceValue[EG][pos.captured_piece()] <= alpha
- || cutNode
- || (!PvNode && !formerPv && captureHistory[movedPiece][to_sq(move)][type_of(pos.captured_piece())] < 3678)
- || thisThread->ttHitAverage < 432 * TtHitAverageResolution * TtHitAverageWindow / 1024))
+ || (cutNode && (ss-1)->moveCount > 1)
+ || (!PvNode && !formerPv))
+ && (!PvNode || ss->ply > 1 || thisThread->id() % 4 != 3))
{
Depth r = reduction(improving, depth, moveCount);
- // Decrease reduction if the ttHit running average is large
+ // Decrease reduction if the ttHit running average is large (~0 Elo)
if (thisThread->ttHitAverage > 537 * TtHitAverageResolution * TtHitAverageWindow / 1024)
r--;
- // Increase reduction if other threads are searching this position
- if (th.marked())
- r++;
-
// Decrease reduction if position is or has been on the PV
- // and node is not likely to fail low. (~10 Elo)
+ // and node is not likely to fail low. (~3 Elo)
if ( ss->ttPv
&& !likelyFailLow)
r -= 2;
&& thisThread->bestMoveChanges <= 2)
r++;
- // More reductions for late moves if position was not in previous PV
- if ( moveCountPruning
- && !formerPv)
- r++;
-
- // Decrease reduction if opponent's move count is high (~5 Elo)
+ // Decrease reduction if opponent's move count is high (~1 Elo)
if ((ss-1)->moveCount > 13)
r--;
- // Decrease reduction if ttMove has been singularly extended (~3 Elo)
+ // Decrease reduction if ttMove has been singularly extended (~1 Elo)
if (singularQuietLMR)
r--;
- if (captureOrPromotion)
- {
- // Increase reduction for non-checking captures likely to be bad
- if ( !givesCheck
- && ss->staticEval + PieceValue[EG][pos.captured_piece()] + 210 * depth <= alpha)
- r++;
- }
- else
+ // Increase reduction for cut nodes (~3 Elo)
+ if (cutNode)
+ r += 1 + !captureOrPromotion;
+
+ if (!captureOrPromotion)
{
- // Increase reduction if ttMove is a capture (~5 Elo)
+ // Increase reduction if ttMove is a capture (~3 Elo)
if (ttCapture)
r++;
- // Increase reduction at root if failing high
- r += rootNode ? thisThread->failedHighCnt * thisThread->failedHighCnt * moveCount / 512 : 0;
-
- // Increase reduction for cut nodes (~10 Elo)
- if (cutNode)
- r += 2;
-
- // Decrease reduction for moves that escape a capture. Filter out
- // castling moves, because they are coded as "king captures rook" and
- // hence break reverse_move() (~2 Elo)
- else if ( type_of(move) == NORMAL
- && !pos.see_ge(reverse_move(move)))
- r -= 2 + ss->ttPv - (type_of(movedPiece) == PAWN);
-
ss->statScore = thisThread->mainHistory[us][from_to(move)]
+ (*contHist[0])[movedPiece][to_sq(move)]
+ (*contHist[1])[movedPiece][to_sq(move)]
+ (*contHist[3])[movedPiece][to_sq(move)]
- - 4741;
-
- // Decrease/increase reduction by comparing opponent's stat score (~10 Elo)
- if (ss->statScore >= -89 && (ss-1)->statScore < -116)
- r--;
-
- else if ((ss-1)->statScore >= -112 && ss->statScore < -100)
- r++;
+ - 4923;
// Decrease/increase reduction for moves with a good/bad history (~30 Elo)
- // If we are not in check use statScore, but if we are in check we use
- // the sum of main history and first continuation history with an offset.
- if (ss->inCheck)
- r -= (thisThread->mainHistory[us][from_to(move)]
- + (*contHist[0])[movedPiece][to_sq(move)] - 3833) / 16384;
- else
- r -= ss->statScore / 14790;
+ if (!ss->inCheck)
+ r -= ss->statScore / 14721;
}
// In general we want to cap the LMR depth search at newDepth. But if
else
{
assert(value >= beta); // Fail high
- ss->statScore = 0;
break;
}
}
Move pv[MAX_PLY+1];
StateInfo st;
- ASSERT_ALIGNED(&st, Eval::NNUE::kCacheLineSize);
+ ASSERT_ALIGNED(&st, Eval::NNUE::CacheLineSize);
TTEntry* tte;
Key posKey;
// and addition of two tempos
ss->staticEval = bestValue =
(ss-1)->currentMove != MOVE_NULL ? evaluate(pos)
- : -(ss-1)->staticEval + 2 * Tempo;
+ : -(ss-1)->staticEval;
// Stand pat. Return immediately if static value is at least beta
if (bestValue >= beta)
[pos.moved_piece(move)]
[to_sq(move)];
- // CounterMove based pruning
+ // Continuation history based pruning
if ( !captureOrPromotion
&& bestValue > VALUE_TB_LOSS_IN_MAX_PLY
&& (*contHist[0])[pos.moved_piece(move)][to_sq(move)] < CounterMovePruneThreshold
bool RootMove::extract_ponder_from_tt(Position& pos) {
StateInfo st;
- ASSERT_ALIGNED(&st, Eval::NNUE::kCacheLineSize);
+ ASSERT_ALIGNED(&st, Eval::NNUE::CacheLineSize);
bool ttHit;