Write text for figure 4.10
This commit is contained in:
@@ -1659,34 +1659,16 @@ sliding-window approach is still at an advantage.
|
||||
|
||||
% [Thread] Exploration of the effect of the step size
|
||||
|
||||
% TODO: Write
|
||||
|
||||
% [Experimental parameters] Figure 4.10
|
||||
|
||||
% tex-fmt: off
|
||||
\red{\textbf{overall:}[warm, cold $F\in\{1,2,3\}$][$W=5$]}
|
||||
\red{\textbf{a)}[$p \in \{\ldots\}$][$n_\text{iter} = 200$]}
|
||||
\red{\textbf{b)}[$p = 0.0025$][$n_\text{iter}\in\{...\}$]}
|
||||
|
||||
% [Description] Figure 4.10
|
||||
|
||||
\red{\textbf{a)}[lower F -> better performance, lower p -> larger
|
||||
gain of warm vs soft, \textbf{TODO}: find more]}
|
||||
\red{\textbf{b)}[lower F -> better performance, lower $n_\text{iter}$
|
||||
-> larger gain of warm vs soft, no real saturation, \textbf{TODO}: find more]}
|
||||
% tex-fmt: on
|
||||
|
||||
% [Interpretation] Figure 4.10
|
||||
|
||||
\red{[lower $n_\text{iter}$ -> larger gain is same behavior as seen
|
||||
in plot before]}
|
||||
\red{[lower F -> better performance makes sense for the same reason
|
||||
larger W -> better performance: greater overlap]}
|
||||
|
||||
% At some later point
|
||||
\content{When looking at max iterations: Callback to diminishing
|
||||
returns with growing window size: More iterations more beneficial
|
||||
than larger window (+1 for warm-start)}
|
||||
Having examined the effect of the window size $W$, we next turned to
|
||||
the second windowing parameter, the step size $F$.
|
||||
We carried out an investigation analogous to the one above:
|
||||
we first compared warm- and cold-start decoding across the full range
|
||||
of physical error rates at a fixed iteration budget, and then we
|
||||
examined the dependence on the iteration budget at a fixed physical
|
||||
error rate.
|
||||
The window size was held fixed at $W = 5$ throughout, the value at
|
||||
which the warm-start variant produced the strongest performance in the
|
||||
previous experiments.
|
||||
|
||||
\begin{figure}[t]
|
||||
\centering
|
||||
@@ -1757,6 +1739,7 @@ than larger window (+1 for warm-start)}
|
||||
\end{tikzpicture}
|
||||
|
||||
\caption{Comparison of window sizes for $F=1$.}
|
||||
\label{fig:bp_f_over_p}
|
||||
\end{subfigure}%
|
||||
\hfill%
|
||||
\begin{subfigure}{0.48\textwidth}
|
||||
@@ -1864,13 +1847,105 @@ than larger window (+1 for warm-start)}
|
||||
\vspace{-3.2mm}
|
||||
|
||||
\caption{Comparison of step sizes for $W=5$.}
|
||||
\label{fig:bp_f_over_iter}
|
||||
\end{subfigure}
|
||||
|
||||
\caption{
|
||||
\red{\lipsum[2]}
|
||||
}
|
||||
\label{fig:bp_f}
|
||||
\end{figure}
|
||||
|
||||
% [Experimental parameters] Figure 4.10
|
||||
|
||||
\Cref{fig:bp_f} summarizes the results of this investigation.
|
||||
|
||||
In both panels the dashed colored curves correspond to cold-start
|
||||
sliding-window decoding for $F \in \{1, 2, 3\}$ and the solid colored
|
||||
curves to the corresponding warm-start runs.
|
||||
The window size is fixed to $W = 5$ throughout.
|
||||
\Cref{fig:bp_f_over_p} sweeps the physical error rate over
|
||||
$p \in [0.001, 0.004]$ in steps of $0.0005$ at a fixed maximum of
|
||||
$n_\text{iter} = 200$ \ac{bp} iterations per window invocation,
|
||||
mirroring the experimental setup of \Cref{fig:whole_vs_cold_vs_warm}.
|
||||
\Cref{fig:bp_f_over_iter} fixes the physical error rate at
|
||||
$p = 0.0025$ and sweeps the iteration budget over
|
||||
$n_\text{iter} \in \{32, 128, 256, 512, 1024, 2048, 4096\}$,
|
||||
mirroring the setup of \Cref{fig:bp_w_over_iter} and again including
|
||||
an inset that magnifies the low-iteration regime
|
||||
$n_\text{iter} \in [32, 512]$.
|
||||
|
||||
% [Description] Figure 4.10
|
||||
|
||||
In \Cref{fig:bp_f_over_p}, every curve exhibits the expected
|
||||
monotonic increase of the per-round \ac{ler} with the physical
|
||||
error rate.
|
||||
At fixed $F$, the warm-start approach lies below
|
||||
cold-start across the entire sweep, and at fixed
|
||||
warm- or cold-start, smaller $F$ produces a lower \ac{ler}.
|
||||
Both gaps grow as the physical error rate decreases:
|
||||
the curves at $F = 1$ separate further from those at $F = 2$ and $F = 3$,
|
||||
and the warm-start curves separate further from the cold-start ones.
|
||||
In \Cref{fig:bp_f_over_iter}, all six curves again decrease
|
||||
monotonically with the iteration budget, with no clear saturation
|
||||
even at $n_\text{iter} = 4096$.
|
||||
Lower $F$ yields a lower \ac{ler} throughout, and warm-start
|
||||
consistently outperforms cold-start at matching $F$.
|
||||
At $n_\text{iter} = 32$, all three cold-start curves coincide at
|
||||
roughly the same per-round \ac{ler}, while the warm-start curves are
|
||||
visibly spread out.
|
||||
Furthermore, the magnified plot confirms that the gap between warm-
|
||||
and cold-start curves at fixed $F$ shrinks as $n_\text{iter}$ grows,
|
||||
and that at fixed $n_\text{iter}$ this gap is largest for $F = 1$.
|
||||
|
||||
% [Interpretation] Figure 4.10
|
||||
|
||||
The observed dependence on the step size mirrors the dependence on
|
||||
the window size studied earlier and the same explanation applies.
|
||||
With $W$ held fixed, decreasing $F$ enlarges the overlap between
|
||||
consecutive windows from $W - F$ to $W - F + 1$ syndrome measurement rounds, so
|
||||
a smaller step size is beneficial for the same reason that a larger
|
||||
window size is:
|
||||
each \ac{vn} in an overlap region participates in more window
|
||||
invocations, and the warm-start modification effectively accumulates
|
||||
iterations on it across these invocations.
|
||||
The widening of the warm/cold gap towards low iteration counts and
|
||||
low physical error rates similarly mirrors the patterns already
|
||||
observed in
|
||||
\Cref{fig:whole_vs_cold_vs_warm,fig:bp_w_over_iter}.
|
||||
|
||||
% TODO: Rephrase
|
||||
The coincidence of all three cold-start curves at
|
||||
$n_\text{iter} = 32$ is a direct consequence of the cold-start initialization.
|
||||
With each new window starting from a uniform prior regardless of $F$,
|
||||
the per-window decoding problem is essentially the same for every
|
||||
step size, and the corresponding \acp{ler} agree as long as the
|
||||
inner decoder has too few iterations to propagate information
|
||||
beyond the local syndrome structure within a single window.
|
||||
This is also the regime in which the warm-start advantage is most
|
||||
valuable, and indeed it is where the warm-start curves spread out
|
||||
most strongly with $F$.
|
||||
|
||||
% TODO: Rephrase
|
||||
A noteworthy methodological point is that, in contrast to the window
|
||||
size $W$, the step size $F$ has no effect on decoding latency:
|
||||
the time at which the inner decoder for a given window can begin
|
||||
running is determined solely by when the syndromes for the rounds
|
||||
covered by that window have been collected, which is independent of
|
||||
how much the window overlaps with its predecessor.
|
||||
A smaller $F$ thus only costs additional total compute and not
|
||||
additional latency, which is favorable for a warm-start
|
||||
sliding-window implementation:
|
||||
the regime in which the warm-start modification helps most --- large
|
||||
overlap and therefore small $F$ --- is precisely the regime in which
|
||||
the cost of that overlap shows up only in the compute budget and not
|
||||
in the latency budget.
|
||||
|
||||
% At some later point
|
||||
\content{When looking at max iterations: Callback to diminishing
|
||||
returns with growing window size: More iterations more beneficial
|
||||
than larger window (+1 for warm-start)}
|
||||
|
||||
%%%%%%%%%%%%%%%%
|
||||
\subsection{Belief Propagation with Guided Decimation}
|
||||
\label{subsec:Belief Propagation with Guided Decimation}
|
||||
|
||||
Reference in New Issue
Block a user