Dynamic Construction of the Lovász Local Lemma

2026-04-22Data Structures and Algorithms

Data Structures and Algorithms
AI summary

The authors show that many local search algorithms, which fix problems step-by-step by resampling or backtracking, still work efficiently even when the problem changes dynamically due to an adaptive adversary adding or removing constraints. They extend previous theoretical tools to prove that the total time spent fixing issues grows almost linearly with updates, instead of exploding. This means these simple and practical algorithms can handle changing problems quickly in a fully dynamic setting. For example, they improve how quickly a certain edge-coloring problem can be maintained compared to prior methods.

Local Search AlgorithmsLovász Local LemmaWitness-tree TechniqueEntropy CompressionLocal Constraint Satisfaction ProblemsBacktrackingLocal ResamplingAdaptive AdversaryDynamic AlgorithmsEdge Coloring
Authors
Bernhard Haeupler, Slobodan Mitrović, Srikkanth Ramachandran, Wen-Horng Sheu, Robert Tarjan
Abstract
This paper proves that a wide class of local search algorithms extend as is to the fully dynamic setting with an adaptive adversary, achieving an amortized $\tilde{O}(1)$ number of local-search steps per update. A breakthrough by Moser (2009) introduced the witness-tree and entropy compression techniques for analyzing local resampling processes for the Lovász Local Lemma. These methods have since been generalized and expanded to analyze a wide variety of local search algorithms that can efficiently find solutions to many important local constraint satisfaction problems. These algorithms either extend a partial valid assignment and backtrack by unassigning variables when constraints become violated, or they iteratively fix violated constraints by resampling their variables. These local resampling or backtracking procedures are incredibly flexible, practical, and simple to specify and implement. Yet, they can be shown to be extremely efficient on static instances, typically performing only (sub)-linear number of fixing steps. The main technical challenge lies in proving conditions that guarantee such rapid convergence. This paper extends these convergence results to fully dynamic settings, where an adaptive adversary may add or remove constraints. We prove that applying the same simple local search procedures to fix old or newly introduced violations leads to a total number of resampling steps near-linear in the number of adversarial updates. Our result is very general and yields several immediate corollaries. For example, letting $Δ$ denote the maximum degree, for a constant $ε$ and $Δ= \text{poly}(\log n)$, we can maintain a $(1+ε) Δ$-edge coloring in $\text{poly}(\log n)$ amortized update time against an adaptive adversary. The prior work for this regime has exponential running time in $\sqrt{\log n}$ [Christiansen, SODA '26].