2 Commits

View File

@@ -1659,34 +1659,16 @@ sliding-window approach is still at an advantage.
% [Thread] Exploration of the effect of the step size
% TODO: Write
% [Experimental parameters] Figure 4.10
% tex-fmt: off
\red{\textbf{overall:}[warm, cold $F\in\{1,2,3\}$][$W=5$]}
\red{\textbf{a)}[$p \in \{\ldots\}$][$n_\text{iter} = 200$]}
\red{\textbf{b)}[$p = 0.0025$][$n_\text{iter}\in\{...\}$]}
% [Description] Figure 4.10
\red{\textbf{a)}[lower F -> better performance, lower p -> larger
gain of warm vs soft, \textbf{TODO}: find more]}
\red{\textbf{b)}[lower F -> better performance, lower $n_\text{iter}$
-> larger gain of warm vs soft, no real saturation, \textbf{TODO}: find more]}
% tex-fmt: on
% [Interpretation] Figure 4.10
\red{[lower $n_\text{iter}$ -> larger gain is same behavior as seen
in plot before]}
\red{[lower F -> better performance makes sense for the same reason
larger W -> better performance: greater overlap]}
% At some later point
\content{When looking at max iterations: Callback to diminishing
returns with growing window size: More iterations more beneficial
than larger window (+1 for warm-start)}
Having examined the effect of the window size $W$, we next turned to
the second windowing parameter, the step size $F$.
We carried out an investigation analogous to the one above:
we first compared warm- and cold-start decoding across the full range
of physical error rates at a fixed iteration budget, and then we
examined the dependence on the iteration budget at a fixed physical
error rate.
The window size was held fixed at $W = 5$ throughout, the value at
which the warm-start variant produced the strongest performance in the
previous experiments.
\begin{figure}[t]
\centering
@@ -1757,6 +1739,7 @@ than larger window (+1 for warm-start)}
\end{tikzpicture}
\caption{Comparison of window sizes for $F=1$.}
\label{fig:bp_f_over_p}
\end{subfigure}%
\hfill%
\begin{subfigure}{0.48\textwidth}
@@ -1864,13 +1847,118 @@ than larger window (+1 for warm-start)}
\vspace{-3.2mm}
\caption{Comparison of step sizes for $W=5$.}
\label{fig:bp_f_over_iter}
\end{subfigure}
\caption{
\red{\lipsum[2]}
}
\label{fig:bp_f}
\end{figure}
% [Experimental parameters] Figure 4.10
\Cref{fig:bp_f} summarizes the results of this investigation.
In both panels the dashed colored curves correspond to cold-start
sliding-window decoding for $F \in \{1, 2, 3\}$ and the solid colored
curves to the corresponding warm-start runs.
The window size is fixed to $W = 5$ throughout.
\Cref{fig:bp_f_over_p} sweeps the physical error rate over
$p \in [0.001, 0.004]$ in steps of $0.0005$ at a fixed maximum of
$n_\text{iter} = 200$ \ac{bp} iterations per window invocation,
mirroring the experimental setup of \Cref{fig:whole_vs_cold_vs_warm}.
\Cref{fig:bp_f_over_iter} fixes the physical error rate at
$p = 0.0025$ and sweeps the iteration budget over
$n_\text{iter} \in \{32, 128, 256, 512, 1024, 2048, 4096\}$,
mirroring the setup of \Cref{fig:bp_w_over_iter} and again including
an inset that magnifies the low-iteration regime
$n_\text{iter} \in [32, 512]$.
% [Description] Figure 4.10
In \Cref{fig:bp_f_over_p}, every curve exhibits the expected
monotonic increase of the per-round \ac{ler} with the physical
error rate.
At fixed $F$, the warm-start approach lies below
cold-start across the entire sweep, and at fixed
warm- or cold-start, smaller $F$ produces a lower \ac{ler}.
Both gaps grow as the physical error rate decreases:
the curves at $F = 1$ separate further from those at $F = 2$ and $F = 3$,
and the warm-start curves separate further from the cold-start ones.
In \Cref{fig:bp_f_over_iter}, all six curves again decrease
monotonically with the iteration budget, with no clear saturation
even at $n_\text{iter} = 4096$.
Lower $F$ yields a lower \ac{ler} throughout, and warm-start
consistently outperforms cold-start at matching $F$.
At $n_\text{iter} = 32$, all three cold-start curves coincide at
roughly the same per-round \ac{ler}, while the warm-start curves are
visibly spread out.
Furthermore, the magnified plot confirms that the gap between warm-
and cold-start curves at fixed $F$ shrinks as $n_\text{iter}$ grows,
and that at fixed $n_\text{iter}$ this gap is largest for $F = 1$.
% [Interpretation] Figure 4.10
The observed dependence on the step size mirrors the dependence on
the window size studied earlier and the same explanation applies.
With $W$ held fixed, decreasing $F$ enlarges the overlap between
consecutive windows from $W - F$ to $W - F + 1$ syndrome measurement rounds, so
a smaller step size is beneficial for the same reason that a larger
window size is:
each \ac{vn} in an overlap region participates in more window
invocations, and the warm-start modification effectively accumulates
iterations on it across these invocations.
The widening of the warm/cold gap towards low iteration counts and
low physical error rates similarly mirrors the patterns already
observed in
\Cref{fig:whole_vs_cold_vs_warm,fig:bp_w_over_iter}.
In contrast to the window size $W$, the step size $F$ has no effect
on decoding latency.
The time at which the inner decoder for a given window can begin
decoding is determined solely by when the syndromes for the rounds
covered by that window have been collected, which is independent of
how much the window overlaps with its predecessor.
Similarly, assuming the decoder is fast enough to keep up with the
incoming syndrome measurements corresponding to the \acp{cn} of
subsequent windows, the time at which decoding is complete depends only
on the amount of time spent on decoding the very last window.
A smaller $F$ thus only costs additional total compute and not
additional latency, which is favorable for a warm-start
sliding-window implementation.
This is especially favorable for our warm-start modification, as it
works best where the overlap is largest, i.e., for low values of $F$.
% Conclusion of BP investigation
We conclude our investigation into the performance of warm-start
sliding-window decoding under plain \ac{bp} by summarizing our findings.
The warm-start modification raises the number of \ac{bp} iterations
effectively spent on the \acp{vn} in an overlap region by reusing the
messages from the previous window invocation instead of restarting
from scratch.
This explains why decoding performance improved monotonically with
the size of the overlap, and consequently why both larger window
sizes $W$ and smaller step sizes $F$ yielded lower per-round \acp{ler}.
The warm-start gain over cold-start was most pronounced at low
per-window iteration budgets,
% and at low physical error rates, the
% regimes
the regime in which each additional iteration carries proportionally
more information.
Additionally, we would like to note that the warm-start modification
incurs no computational cost relative to cold-start decoding.
It changes neither the decoding latency nor the total compute, since
both schemes process the same windows for the same number of
iterations and differ only in the initialization of the \ac{bp}
messages of each new window.
We also observed that plain \ac{bp} did not saturate even at $4096$
iterations, which we attribute to the short cycles in the underlying
Tanner graph.
This motivates the next subsection, in which we replace the inner
\ac{bp} decoder by its guided-decimation variant.
%%%%%%%%%%%%%%%%
\subsection{Belief Propagation with Guided Decimation}
\label{subsec:Belief Propagation with Guided Decimation}