Write text for figure 4.10
This commit is contained in:
@@ -1659,34 +1659,16 @@ sliding-window approach is still at an advantage.
|
|||||||
|
|
||||||
% [Thread] Exploration of the effect of the step size
|
% [Thread] Exploration of the effect of the step size
|
||||||
|
|
||||||
% TODO: Write
|
Having examined the effect of the window size $W$, we next turned to
|
||||||
|
the second windowing parameter, the step size $F$.
|
||||||
% [Experimental parameters] Figure 4.10
|
We carried out an investigation analogous to the one above:
|
||||||
|
we first compared warm- and cold-start decoding across the full range
|
||||||
% tex-fmt: off
|
of physical error rates at a fixed iteration budget, and then we
|
||||||
\red{\textbf{overall:}[warm, cold $F\in\{1,2,3\}$][$W=5$]}
|
examined the dependence on the iteration budget at a fixed physical
|
||||||
\red{\textbf{a)}[$p \in \{\ldots\}$][$n_\text{iter} = 200$]}
|
error rate.
|
||||||
\red{\textbf{b)}[$p = 0.0025$][$n_\text{iter}\in\{...\}$]}
|
The window size was held fixed at $W = 5$ throughout, the value at
|
||||||
|
which the warm-start variant produced the strongest performance in the
|
||||||
% [Description] Figure 4.10
|
previous experiments.
|
||||||
|
|
||||||
\red{\textbf{a)}[lower F -> better performance, lower p -> larger
|
|
||||||
gain of warm vs soft, \textbf{TODO}: find more]}
|
|
||||||
\red{\textbf{b)}[lower F -> better performance, lower $n_\text{iter}$
|
|
||||||
-> larger gain of warm vs soft, no real saturation, \textbf{TODO}: find more]}
|
|
||||||
% tex-fmt: on
|
|
||||||
|
|
||||||
% [Interpretation] Figure 4.10
|
|
||||||
|
|
||||||
\red{[lower $n_\text{iter}$ -> larger gain is same behavior as seen
|
|
||||||
in plot before]}
|
|
||||||
\red{[lower F -> better performance makes sense for the same reason
|
|
||||||
larger W -> better performance: greater overlap]}
|
|
||||||
|
|
||||||
% At some later point
|
|
||||||
\content{When looking at max iterations: Callback to diminishing
|
|
||||||
returns with growing window size: More iterations more beneficial
|
|
||||||
than larger window (+1 for warm-start)}
|
|
||||||
|
|
||||||
\begin{figure}[t]
|
\begin{figure}[t]
|
||||||
\centering
|
\centering
|
||||||
@@ -1757,6 +1739,7 @@ than larger window (+1 for warm-start)}
|
|||||||
\end{tikzpicture}
|
\end{tikzpicture}
|
||||||
|
|
||||||
\caption{Comparison of window sizes for $F=1$.}
|
\caption{Comparison of window sizes for $F=1$.}
|
||||||
|
\label{fig:bp_f_over_p}
|
||||||
\end{subfigure}%
|
\end{subfigure}%
|
||||||
\hfill%
|
\hfill%
|
||||||
\begin{subfigure}{0.48\textwidth}
|
\begin{subfigure}{0.48\textwidth}
|
||||||
@@ -1864,13 +1847,105 @@ than larger window (+1 for warm-start)}
|
|||||||
\vspace{-3.2mm}
|
\vspace{-3.2mm}
|
||||||
|
|
||||||
\caption{Comparison of step sizes for $W=5$.}
|
\caption{Comparison of step sizes for $W=5$.}
|
||||||
|
\label{fig:bp_f_over_iter}
|
||||||
\end{subfigure}
|
\end{subfigure}
|
||||||
|
|
||||||
\caption{
|
\caption{
|
||||||
\red{\lipsum[2]}
|
\red{\lipsum[2]}
|
||||||
}
|
}
|
||||||
|
\label{fig:bp_f}
|
||||||
\end{figure}
|
\end{figure}
|
||||||
|
|
||||||
|
% [Experimental parameters] Figure 4.10
|
||||||
|
|
||||||
|
\Cref{fig:bp_f} summarizes the results of this investigation.
|
||||||
|
|
||||||
|
In both panels the dashed colored curves correspond to cold-start
|
||||||
|
sliding-window decoding for $F \in \{1, 2, 3\}$ and the solid colored
|
||||||
|
curves to the corresponding warm-start runs.
|
||||||
|
The window size is fixed to $W = 5$ throughout.
|
||||||
|
\Cref{fig:bp_f_over_p} sweeps the physical error rate over
|
||||||
|
$p \in [0.001, 0.004]$ in steps of $0.0005$ at a fixed maximum of
|
||||||
|
$n_\text{iter} = 200$ \ac{bp} iterations per window invocation,
|
||||||
|
mirroring the experimental setup of \Cref{fig:whole_vs_cold_vs_warm}.
|
||||||
|
\Cref{fig:bp_f_over_iter} fixes the physical error rate at
|
||||||
|
$p = 0.0025$ and sweeps the iteration budget over
|
||||||
|
$n_\text{iter} \in \{32, 128, 256, 512, 1024, 2048, 4096\}$,
|
||||||
|
mirroring the setup of \Cref{fig:bp_w_over_iter} and again including
|
||||||
|
an inset that magnifies the low-iteration regime
|
||||||
|
$n_\text{iter} \in [32, 512]$.
|
||||||
|
|
||||||
|
% [Description] Figure 4.10
|
||||||
|
|
||||||
|
In \Cref{fig:bp_f_over_p}, every curve exhibits the expected
|
||||||
|
monotonic increase of the per-round \ac{ler} with the physical
|
||||||
|
error rate.
|
||||||
|
At fixed $F$, the warm-start approach lies below
|
||||||
|
cold-start across the entire sweep, and at fixed
|
||||||
|
warm- or cold-start, smaller $F$ produces a lower \ac{ler}.
|
||||||
|
Both gaps grow as the physical error rate decreases:
|
||||||
|
the curves at $F = 1$ separate further from those at $F = 2$ and $F = 3$,
|
||||||
|
and the warm-start curves separate further from the cold-start ones.
|
||||||
|
In \Cref{fig:bp_f_over_iter}, all six curves again decrease
|
||||||
|
monotonically with the iteration budget, with no clear saturation
|
||||||
|
even at $n_\text{iter} = 4096$.
|
||||||
|
Lower $F$ yields a lower \ac{ler} throughout, and warm-start
|
||||||
|
consistently outperforms cold-start at matching $F$.
|
||||||
|
At $n_\text{iter} = 32$, all three cold-start curves coincide at
|
||||||
|
roughly the same per-round \ac{ler}, while the warm-start curves are
|
||||||
|
visibly spread out.
|
||||||
|
Furthermore, the magnified plot confirms that the gap between warm-
|
||||||
|
and cold-start curves at fixed $F$ shrinks as $n_\text{iter}$ grows,
|
||||||
|
and that at fixed $n_\text{iter}$ this gap is largest for $F = 1$.
|
||||||
|
|
||||||
|
% [Interpretation] Figure 4.10
|
||||||
|
|
||||||
|
The observed dependence on the step size mirrors the dependence on
|
||||||
|
the window size studied earlier and the same explanation applies.
|
||||||
|
With $W$ held fixed, decreasing $F$ enlarges the overlap between
|
||||||
|
consecutive windows from $W - F$ to $W - F + 1$ syndrome measurement rounds, so
|
||||||
|
a smaller step size is beneficial for the same reason that a larger
|
||||||
|
window size is:
|
||||||
|
each \ac{vn} in an overlap region participates in more window
|
||||||
|
invocations, and the warm-start modification effectively accumulates
|
||||||
|
iterations on it across these invocations.
|
||||||
|
The widening of the warm/cold gap towards low iteration counts and
|
||||||
|
low physical error rates similarly mirrors the patterns already
|
||||||
|
observed in
|
||||||
|
\Cref{fig:whole_vs_cold_vs_warm,fig:bp_w_over_iter}.
|
||||||
|
|
||||||
|
% TODO: Rephrase
|
||||||
|
The coincidence of all three cold-start curves at
|
||||||
|
$n_\text{iter} = 32$ is a direct consequence of the cold-start initialization.
|
||||||
|
With each new window starting from a uniform prior regardless of $F$,
|
||||||
|
the per-window decoding problem is essentially the same for every
|
||||||
|
step size, and the corresponding \acp{ler} agree as long as the
|
||||||
|
inner decoder has too few iterations to propagate information
|
||||||
|
beyond the local syndrome structure within a single window.
|
||||||
|
This is also the regime in which the warm-start advantage is most
|
||||||
|
valuable, and indeed it is where the warm-start curves spread out
|
||||||
|
most strongly with $F$.
|
||||||
|
|
||||||
|
% TODO: Rephrase
|
||||||
|
A noteworthy methodological point is that, in contrast to the window
|
||||||
|
size $W$, the step size $F$ has no effect on decoding latency:
|
||||||
|
the time at which the inner decoder for a given window can begin
|
||||||
|
running is determined solely by when the syndromes for the rounds
|
||||||
|
covered by that window have been collected, which is independent of
|
||||||
|
how much the window overlaps with its predecessor.
|
||||||
|
A smaller $F$ thus only costs additional total compute and not
|
||||||
|
additional latency, which is favorable for a warm-start
|
||||||
|
sliding-window implementation:
|
||||||
|
the regime in which the warm-start modification helps most --- large
|
||||||
|
overlap and therefore small $F$ --- is precisely the regime in which
|
||||||
|
the cost of that overlap shows up only in the compute budget and not
|
||||||
|
in the latency budget.
|
||||||
|
|
||||||
|
% At some later point
|
||||||
|
\content{When looking at max iterations: Callback to diminishing
|
||||||
|
returns with growing window size: More iterations more beneficial
|
||||||
|
than larger window (+1 for warm-start)}
|
||||||
|
|
||||||
%%%%%%%%%%%%%%%%
|
%%%%%%%%%%%%%%%%
|
||||||
\subsection{Belief Propagation with Guided Decimation}
|
\subsection{Belief Propagation with Guided Decimation}
|
||||||
\label{subsec:Belief Propagation with Guided Decimation}
|
\label{subsec:Belief Propagation with Guided Decimation}
|
||||||
|
|||||||
Reference in New Issue
Block a user