Write text for figure 4.10

This commit is contained in:
2026-05-02 17:59:44 +02:00
parent 6b1821fd6b
commit d2960b8f0e

View File

@@ -1659,34 +1659,16 @@ sliding-window approach is still at an advantage.
% [Thread] Exploration of the effect of the step size
% TODO: Write
% [Experimental parameters] Figure 4.10
% tex-fmt: off
\red{\textbf{overall:}[warm, cold $F\in\{1,2,3\}$][$W=5$]}
\red{\textbf{a)}[$p \in \{\ldots\}$][$n_\text{iter} = 200$]}
\red{\textbf{b)}[$p = 0.0025$][$n_\text{iter}\in\{...\}$]}
% [Description] Figure 4.10
\red{\textbf{a)}[lower F -> better performance, lower p -> larger
gain of warm vs soft, \textbf{TODO}: find more]}
\red{\textbf{b)}[lower F -> better performance, lower $n_\text{iter}$
-> larger gain of warm vs soft, no real saturation, \textbf{TODO}: find more]}
% tex-fmt: on
% [Interpretation] Figure 4.10
\red{[lower $n_\text{iter}$ -> larger gain is same behavior as seen
in plot before]}
\red{[lower F -> better performance makes sense for the same reason
larger W -> better performance: greater overlap]}
% At some later point
\content{When looking at max iterations: Callback to diminishing
returns with growing window size: More iterations more beneficial
than larger window (+1 for warm-start)}
Having examined the effect of the window size $W$, we next turned to
the second windowing parameter, the step size $F$.
We carried out an investigation analogous to the one above:
we first compared warm- and cold-start decoding across the full range
of physical error rates at a fixed iteration budget, and then we
examined the dependence on the iteration budget at a fixed physical
error rate.
The window size was held fixed at $W = 5$ throughout, the value at
which the warm-start variant produced the strongest performance in the
previous experiments.
\begin{figure}[t]
\centering
@@ -1757,6 +1739,7 @@ than larger window (+1 for warm-start)}
\end{tikzpicture}
\caption{Comparison of window sizes for $F=1$.}
\label{fig:bp_f_over_p}
\end{subfigure}%
\hfill%
\begin{subfigure}{0.48\textwidth}
@@ -1864,13 +1847,105 @@ than larger window (+1 for warm-start)}
\vspace{-3.2mm}
\caption{Comparison of step sizes for $W=5$.}
\label{fig:bp_f_over_iter}
\end{subfigure}
\caption{
\red{\lipsum[2]}
}
\label{fig:bp_f}
\end{figure}
% [Experimental parameters] Figure 4.10
\Cref{fig:bp_f} summarizes the results of this investigation.
In both panels the dashed colored curves correspond to cold-start
sliding-window decoding for $F \in \{1, 2, 3\}$ and the solid colored
curves to the corresponding warm-start runs.
The window size is fixed to $W = 5$ throughout.
\Cref{fig:bp_f_over_p} sweeps the physical error rate over
$p \in [0.001, 0.004]$ in steps of $0.0005$ at a fixed maximum of
$n_\text{iter} = 200$ \ac{bp} iterations per window invocation,
mirroring the experimental setup of \Cref{fig:whole_vs_cold_vs_warm}.
\Cref{fig:bp_f_over_iter} fixes the physical error rate at
$p = 0.0025$ and sweeps the iteration budget over
$n_\text{iter} \in \{32, 128, 256, 512, 1024, 2048, 4096\}$,
mirroring the setup of \Cref{fig:bp_w_over_iter} and again including
an inset that magnifies the low-iteration regime
$n_\text{iter} \in [32, 512]$.
% [Description] Figure 4.10
In \Cref{fig:bp_f_over_p}, every curve exhibits the expected
monotonic increase of the per-round \ac{ler} with the physical
error rate.
At fixed $F$, the warm-start approach lies below
cold-start across the entire sweep, and at fixed
warm- or cold-start, smaller $F$ produces a lower \ac{ler}.
Both gaps grow as the physical error rate decreases:
the curves at $F = 1$ separate further from those at $F = 2$ and $F = 3$,
and the warm-start curves separate further from the cold-start ones.
In \Cref{fig:bp_f_over_iter}, all six curves again decrease
monotonically with the iteration budget, with no clear saturation
even at $n_\text{iter} = 4096$.
Lower $F$ yields a lower \ac{ler} throughout, and warm-start
consistently outperforms cold-start at matching $F$.
At $n_\text{iter} = 32$, all three cold-start curves coincide at
roughly the same per-round \ac{ler}, while the warm-start curves are
visibly spread out.
Furthermore, the magnified plot confirms that the gap between warm-
and cold-start curves at fixed $F$ shrinks as $n_\text{iter}$ grows,
and that at fixed $n_\text{iter}$ this gap is largest for $F = 1$.
% [Interpretation] Figure 4.10
The observed dependence on the step size mirrors the dependence on
the window size studied earlier and the same explanation applies.
With $W$ held fixed, decreasing $F$ enlarges the overlap between
consecutive windows from $W - F$ to $W - F + 1$ syndrome measurement rounds, so
a smaller step size is beneficial for the same reason that a larger
window size is:
each \ac{vn} in an overlap region participates in more window
invocations, and the warm-start modification effectively accumulates
iterations on it across these invocations.
The widening of the warm/cold gap towards low iteration counts and
low physical error rates similarly mirrors the patterns already
observed in
\Cref{fig:whole_vs_cold_vs_warm,fig:bp_w_over_iter}.
% TODO: Rephrase
The coincidence of all three cold-start curves at
$n_\text{iter} = 32$ is a direct consequence of the cold-start initialization.
With each new window starting from a uniform prior regardless of $F$,
the per-window decoding problem is essentially the same for every
step size, and the corresponding \acp{ler} agree as long as the
inner decoder has too few iterations to propagate information
beyond the local syndrome structure within a single window.
This is also the regime in which the warm-start advantage is most
valuable, and indeed it is where the warm-start curves spread out
most strongly with $F$.
% TODO: Rephrase
A noteworthy methodological point is that, in contrast to the window
size $W$, the step size $F$ has no effect on decoding latency:
the time at which the inner decoder for a given window can begin
running is determined solely by when the syndromes for the rounds
covered by that window have been collected, which is independent of
how much the window overlaps with its predecessor.
A smaller $F$ thus only costs additional total compute and not
additional latency, which is favorable for a warm-start
sliding-window implementation:
the regime in which the warm-start modification helps most --- large
overlap and therefore small $F$ --- is precisely the regime in which
the cost of that overlap shows up only in the compute budget and not
in the latency budget.
% At some later point
\content{When looking at max iterations: Callback to diminishing
returns with growing window size: More iterations more beneficial
than larger window (+1 for warm-start)}
%%%%%%%%%%%%%%%%
\subsection{Belief Propagation with Guided Decimation}
\label{subsec:Belief Propagation with Guided Decimation}