Compare commits
2 Commits
6b1821fd6b
...
a90458dd8a
| Author | SHA1 | Date | |
|---|---|---|---|
| a90458dd8a | |||
| d2960b8f0e |
@@ -1659,34 +1659,16 @@ sliding-window approach is still at an advantage.
|
||||
|
||||
% [Thread] Exploration of the effect of the step size
|
||||
|
||||
% TODO: Write
|
||||
|
||||
% [Experimental parameters] Figure 4.10
|
||||
|
||||
% tex-fmt: off
|
||||
\red{\textbf{overall:}[warm, cold $F\in\{1,2,3\}$][$W=5$]}
|
||||
\red{\textbf{a)}[$p \in \{\ldots\}$][$n_\text{iter} = 200$]}
|
||||
\red{\textbf{b)}[$p = 0.0025$][$n_\text{iter}\in\{...\}$]}
|
||||
|
||||
% [Description] Figure 4.10
|
||||
|
||||
\red{\textbf{a)}[lower F -> better performance, lower p -> larger
|
||||
gain of warm vs soft, \textbf{TODO}: find more]}
|
||||
\red{\textbf{b)}[lower F -> better performance, lower $n_\text{iter}$
|
||||
-> larger gain of warm vs soft, no real saturation, \textbf{TODO}: find more]}
|
||||
% tex-fmt: on
|
||||
|
||||
% [Interpretation] Figure 4.10
|
||||
|
||||
\red{[lower $n_\text{iter}$ -> larger gain is same behavior as seen
|
||||
in plot before]}
|
||||
\red{[lower F -> better performance makes sense for the same reason
|
||||
larger W -> better performance: greater overlap]}
|
||||
|
||||
% At some later point
|
||||
\content{When looking at max iterations: Callback to diminishing
|
||||
returns with growing window size: More iterations more beneficial
|
||||
than larger window (+1 for warm-start)}
|
||||
Having examined the effect of the window size $W$, we next turned to
|
||||
the second windowing parameter, the step size $F$.
|
||||
We carried out an investigation analogous to the one above:
|
||||
we first compared warm- and cold-start decoding across the full range
|
||||
of physical error rates at a fixed iteration budget, and then we
|
||||
examined the dependence on the iteration budget at a fixed physical
|
||||
error rate.
|
||||
The window size was held fixed at $W = 5$ throughout, the value at
|
||||
which the warm-start variant produced the strongest performance in the
|
||||
previous experiments.
|
||||
|
||||
\begin{figure}[t]
|
||||
\centering
|
||||
@@ -1757,6 +1739,7 @@ than larger window (+1 for warm-start)}
|
||||
\end{tikzpicture}
|
||||
|
||||
\caption{Comparison of window sizes for $F=1$.}
|
||||
\label{fig:bp_f_over_p}
|
||||
\end{subfigure}%
|
||||
\hfill%
|
||||
\begin{subfigure}{0.48\textwidth}
|
||||
@@ -1864,13 +1847,118 @@ than larger window (+1 for warm-start)}
|
||||
\vspace{-3.2mm}
|
||||
|
||||
\caption{Comparison of step sizes for $W=5$.}
|
||||
\label{fig:bp_f_over_iter}
|
||||
\end{subfigure}
|
||||
|
||||
\caption{
|
||||
\red{\lipsum[2]}
|
||||
}
|
||||
\label{fig:bp_f}
|
||||
\end{figure}
|
||||
|
||||
% [Experimental parameters] Figure 4.10
|
||||
|
||||
\Cref{fig:bp_f} summarizes the results of this investigation.
|
||||
|
||||
In both panels the dashed colored curves correspond to cold-start
|
||||
sliding-window decoding for $F \in \{1, 2, 3\}$ and the solid colored
|
||||
curves to the corresponding warm-start runs.
|
||||
The window size is fixed to $W = 5$ throughout.
|
||||
\Cref{fig:bp_f_over_p} sweeps the physical error rate over
|
||||
$p \in [0.001, 0.004]$ in steps of $0.0005$ at a fixed maximum of
|
||||
$n_\text{iter} = 200$ \ac{bp} iterations per window invocation,
|
||||
mirroring the experimental setup of \Cref{fig:whole_vs_cold_vs_warm}.
|
||||
\Cref{fig:bp_f_over_iter} fixes the physical error rate at
|
||||
$p = 0.0025$ and sweeps the iteration budget over
|
||||
$n_\text{iter} \in \{32, 128, 256, 512, 1024, 2048, 4096\}$,
|
||||
mirroring the setup of \Cref{fig:bp_w_over_iter} and again including
|
||||
an inset that magnifies the low-iteration regime
|
||||
$n_\text{iter} \in [32, 512]$.
|
||||
|
||||
% [Description] Figure 4.10
|
||||
|
||||
In \Cref{fig:bp_f_over_p}, every curve exhibits the expected
|
||||
monotonic increase of the per-round \ac{ler} with the physical
|
||||
error rate.
|
||||
At fixed $F$, the warm-start approach lies below
|
||||
cold-start across the entire sweep, and at fixed
|
||||
warm- or cold-start, smaller $F$ produces a lower \ac{ler}.
|
||||
Both gaps grow as the physical error rate decreases:
|
||||
the curves at $F = 1$ separate further from those at $F = 2$ and $F = 3$,
|
||||
and the warm-start curves separate further from the cold-start ones.
|
||||
In \Cref{fig:bp_f_over_iter}, all six curves again decrease
|
||||
monotonically with the iteration budget, with no clear saturation
|
||||
even at $n_\text{iter} = 4096$.
|
||||
Lower $F$ yields a lower \ac{ler} throughout, and warm-start
|
||||
consistently outperforms cold-start at matching $F$.
|
||||
At $n_\text{iter} = 32$, all three cold-start curves coincide at
|
||||
roughly the same per-round \ac{ler}, while the warm-start curves are
|
||||
visibly spread out.
|
||||
Furthermore, the magnified plot confirms that the gap between warm-
|
||||
and cold-start curves at fixed $F$ shrinks as $n_\text{iter}$ grows,
|
||||
and that at fixed $n_\text{iter}$ this gap is largest for $F = 1$.
|
||||
|
||||
% [Interpretation] Figure 4.10
|
||||
|
||||
The observed dependence on the step size mirrors the dependence on
|
||||
the window size studied earlier and the same explanation applies.
|
||||
With $W$ held fixed, decreasing $F$ enlarges the overlap between
|
||||
consecutive windows from $W - F$ to $W - F + 1$ syndrome measurement rounds, so
|
||||
a smaller step size is beneficial for the same reason that a larger
|
||||
window size is:
|
||||
each \ac{vn} in an overlap region participates in more window
|
||||
invocations, and the warm-start modification effectively accumulates
|
||||
iterations on it across these invocations.
|
||||
The widening of the warm/cold gap towards low iteration counts and
|
||||
low physical error rates similarly mirrors the patterns already
|
||||
observed in
|
||||
\Cref{fig:whole_vs_cold_vs_warm,fig:bp_w_over_iter}.
|
||||
|
||||
In contrast to the window size $W$, the step size $F$ has no effect
|
||||
on decoding latency.
|
||||
The time at which the inner decoder for a given window can begin
|
||||
decoding is determined solely by when the syndromes for the rounds
|
||||
covered by that window have been collected, which is independent of
|
||||
how much the window overlaps with its predecessor.
|
||||
Similarly, assuming the decoder is fast enough to keep up with the
|
||||
incoming syndrome measurements corresponding to the \acp{cn} of
|
||||
subsequent windows, the time at which decoding is complete depends only
|
||||
on the amount of time spent on decoding the very last window.
|
||||
A smaller $F$ thus only costs additional total compute and not
|
||||
additional latency, which is favorable for a warm-start
|
||||
sliding-window implementation.
|
||||
This is especially favorable for our warm-start modification, as it
|
||||
works best where the overlap is largest, i.e., for low values of $F$.
|
||||
|
||||
% Conclusion of BP investigation
|
||||
|
||||
We conclude our investigation into the performance of warm-start
|
||||
sliding-window decoding under plain \ac{bp} by summarizing our findings.
|
||||
The warm-start modification raises the number of \ac{bp} iterations
|
||||
effectively spent on the \acp{vn} in an overlap region by reusing the
|
||||
messages from the previous window invocation instead of restarting
|
||||
from scratch.
|
||||
This explains why decoding performance improved monotonically with
|
||||
the size of the overlap, and consequently why both larger window
|
||||
sizes $W$ and smaller step sizes $F$ yielded lower per-round \acp{ler}.
|
||||
The warm-start gain over cold-start was most pronounced at low
|
||||
per-window iteration budgets,
|
||||
% and at low physical error rates, the
|
||||
% regimes
|
||||
the regime in which each additional iteration carries proportionally
|
||||
more information.
|
||||
Additionally, we would like to note that the warm-start modification
|
||||
incurs no computational cost relative to cold-start decoding.
|
||||
It changes neither the decoding latency nor the total compute, since
|
||||
both schemes process the same windows for the same number of
|
||||
iterations and differ only in the initialization of the \ac{bp}
|
||||
messages of each new window.
|
||||
We also observed that plain \ac{bp} did not saturate even at $4096$
|
||||
iterations, which we attribute to the short cycles in the underlying
|
||||
Tanner graph.
|
||||
This motivates the next subsection, in which we replace the inner
|
||||
\ac{bp} decoder by its guided-decimation variant.
|
||||
|
||||
%%%%%%%%%%%%%%%%
|
||||
\subsection{Belief Propagation with Guided Decimation}
|
||||
\label{subsec:Belief Propagation with Guided Decimation}
|
||||
|
||||
Reference in New Issue
Block a user