Added text for proximal K and w choice; Wording changes
This commit is contained in:
parent
ea00e6dda4
commit
545fed15c8
@ -140,8 +140,12 @@ descent:%
|
|||||||
.\end{align}%
|
.\end{align}%
|
||||||
%
|
%
|
||||||
For the second step, minimizing the scaled code-constraint polynomial, the
|
For the second step, minimizing the scaled code-constraint polynomial, the
|
||||||
proximal gradient method is used and the \textit{proximal operator} of
|
proximal gradient method is used \todo{The proximal gradient method is not
|
||||||
|
just used for the second step. It is the name for the alternating iterative process}
|
||||||
|
and the \textit{proximal operator} of
|
||||||
$\gamma h\left( \tilde{\boldsymbol{x}} \right) $ has to be computed.
|
$\gamma h\left( \tilde{\boldsymbol{x}} \right) $ has to be computed.
|
||||||
|
\todo{Note about how the proximal gradient method is meant for convex optimization
|
||||||
|
problems but used for a non-convex problem in this case?}
|
||||||
It is then immediately approximated with gradient-descent:%
|
It is then immediately approximated with gradient-descent:%
|
||||||
%
|
%
|
||||||
\begin{align*}
|
\begin{align*}
|
||||||
@ -304,7 +308,7 @@ the gradient can be written as%
|
|||||||
%
|
%
|
||||||
enabling its computation primarily with element-wise operations and
|
enabling its computation primarily with element-wise operations and
|
||||||
matrix-vector multiplication.
|
matrix-vector multiplication.
|
||||||
This is beneficial, as the libraries used for the implementation are
|
This is beneficial, as the libraries employed for the implementation are
|
||||||
heavily optimized for such calculations (e.g., through vectorization of the
|
heavily optimized for such calculations (e.g., through vectorization of the
|
||||||
operations).
|
operations).
|
||||||
\todo{Note about how the equation with which the gradient is calculated is
|
\todo{Note about how the equation with which the gradient is calculated is
|
||||||
@ -430,7 +434,8 @@ Evidently, while the decoding performance does depend on the value of
|
|||||||
$\gamma$, there is no single optimal value offering optimal performance, but
|
$\gamma$, there is no single optimal value offering optimal performance, but
|
||||||
rather a certain interval in which it stays largely unchanged.
|
rather a certain interval in which it stays largely unchanged.
|
||||||
When examining a number of different codes (figure
|
When examining a number of different codes (figure
|
||||||
\ref{fig:prox:results_3d_multiple}), it is apparent that while the exact
|
\ref{fig:prox:results_3d_multiple}), \todo{Move figure to appendix?}
|
||||||
|
it is apparent that while the exact
|
||||||
landscape of the graph depends on the code, the general behaviour is the same
|
landscape of the graph depends on the code, the general behaviour is the same
|
||||||
in each case.
|
in each case.
|
||||||
|
|
||||||
@ -483,25 +488,59 @@ in each case.
|
|||||||
\cite[\text{204.33.484}]{mackay_enc}; $\omega = 0.05, K=200, \eta=1.5$
|
\cite[\text{204.33.484}]{mackay_enc}; $\omega = 0.05, K=200, \eta=1.5$
|
||||||
}%
|
}%
|
||||||
%
|
%
|
||||||
\noindent This indicates \todo{This is a result fit for the conclusion}
|
\noindent This indicates that while the choice of the parameter $\gamma$
|
||||||
that while the choice of the parameter $\gamma$ significantly
|
significantly affects the decoding performance, there is not much benefit
|
||||||
affects the decoding performance, there is not much benefit attainable in
|
attainable in undertaking an extensive search for an exact optimum.
|
||||||
undertaking an extensive search for an exact optimum.
|
|
||||||
Rather, a preliminary examination providing a rough window for $\gamma$ may
|
Rather, a preliminary examination providing a rough window for $\gamma$ may
|
||||||
be sufficient.
|
be sufficient.
|
||||||
|
|
||||||
TODO: $\omega, K$
|
The parameter $\gamma$ describes the step-size for the optimization step
|
||||||
|
dealing with the code-constraint polynomial;
|
||||||
|
the parameter $\omega$ describes the step-size for the step dealing with the
|
||||||
|
negative-log likelihood.
|
||||||
|
The relationship between $\omega$ and $\gamma$ is studied in figure
|
||||||
|
\ref{TODO}.
|
||||||
|
The \ac{SNR} is kept constant at $\SI{4}{dB}$.
|
||||||
|
Similar behaviour to $\gamma$ is exhibited: the \ac{BER} is minimized when
|
||||||
|
keeping the value within certain bounds, without displaying a clear
|
||||||
|
optimum.
|
||||||
|
It is noteworthy that the decoder seems to achieve the best performance for
|
||||||
|
similar values of the two step sizes.
|
||||||
|
Again, this consideration applies to a multitude of different codes, depicted
|
||||||
|
in figure \ref{TODO}.
|
||||||
|
|
||||||
|
To better understand how to determine the optimal value for the parameter $K$,
|
||||||
|
the average error is inspected.
|
||||||
|
This time $\gamma$ and $\omega$ are held constant and the average error is
|
||||||
|
observed during each iteration of the decoding process for a number of
|
||||||
|
different \acp{SNR}.
|
||||||
|
The plots have been generated by averaging the error over TODO decodings.
|
||||||
|
As some decodings go one for more iterations than others, the number of values
|
||||||
|
which are averaged for each datapoints vary.
|
||||||
|
This explains the bump observable around $k=\text{TODO}$, since after
|
||||||
|
this point more and more correct decodings converge and stop iterating,
|
||||||
|
leaving more and more faulty ones to be averaged.
|
||||||
|
Remarkably, the \ac{SNR} seems to not have any impact on the number of
|
||||||
|
iterations necessary to reach the point at which the average error
|
||||||
|
stabilizes.
|
||||||
|
Furthermore, the improvement in decoding performance stagnates at a particular
|
||||||
|
point, rendering an increase in $K$ counterproductive as it only raises the
|
||||||
|
average timing requirements of the decoding process.
|
||||||
|
|
||||||
Changing the parameter $\eta$ does not appear to have a significant effect on
|
Changing the parameter $\eta$ does not appear to have a significant effect on
|
||||||
the decoding performance when keeping the value within a reasonable window
|
the decoding performance when keeping the value within a reasonable window
|
||||||
(''slightly larger than one``, as stated in \cite[Sec. 3.2]{proximal_paper}),
|
(``slightly larger than one'', as stated in \cite[Sec. 3.2]{proximal_paper}),
|
||||||
which seems plausible considering its only function is ensuring numerical stability.
|
which seems plausible considering its only function is ensuring numerical stability.
|
||||||
|
|
||||||
Summarizing the above considerations, \ldots
|
Summarizing the above considerations, an intricate strategy to find the exact
|
||||||
|
optimum values for the parameters $\gamma$ and $\omega$ appears to bring
|
||||||
\begin{itemize}
|
limited benefit;
|
||||||
\item Conclusion: Number of iterations independent of \ac{SNR}
|
an initial rudimentary examination to find the general bounds in which the two
|
||||||
\end{itemize}
|
values should lie is sufficient.
|
||||||
|
The parameter $K$ is independent of the $SNR$ and raising its value above a
|
||||||
|
certain threshold does not improve the decoding performance.
|
||||||
|
The choice of $\eta$ is insignificant and the parameter is only relevant as a
|
||||||
|
means to bring about numerical stability.
|
||||||
|
|
||||||
\begin{figure}[H]
|
\begin{figure}[H]
|
||||||
\centering
|
\centering
|
||||||
@ -1272,6 +1311,7 @@ used due to memory requirements?}
|
|||||||
Some deviations from linear behaviour are unavoidable because not all codes
|
Some deviations from linear behaviour are unavoidable because not all codes
|
||||||
considered are actually \ac{LDPC} codes, or \ac{LDPC} codes constructed
|
considered are actually \ac{LDPC} codes, or \ac{LDPC} codes constructed
|
||||||
according to the same scheme.
|
according to the same scheme.
|
||||||
|
\todo{Mention on what hardware the results where generated}
|
||||||
Nontheless, a generally linear relationship between the average time needed to
|
Nontheless, a generally linear relationship between the average time needed to
|
||||||
decode a received frame and the length $n$ of the frame can be observed.
|
decode a received frame and the length $n$ of the frame can be observed.
|
||||||
|
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user