Efficient Floating-Point Implementation of Signal Processing Algorithms on Reconfigurable Hardware
This doctoral thesis aims at optimising the ?oating-point implementations of signal processing algorithms on recon?gurable hardware with respect to accuracy, hardware resource and execution time. It is known that reduced precision in ?oating-point arithmetic operations on recon?gurable hardware directly translates into increased parallelism and peak performance. As a result, e?cient implementations can be obtained by choosing the minimal acceptable precision for ?oating-point operations. Furthermore, custom-precision ?oating-point operations allow for trading accuracy with parallelism and performance. We use A?ne Arithmetic (AA) for modeling the rounding errors of ?oating-point computations. The derived rounding error bound by the AA-based error model is then used to determine the smallest mantissa bit width of custom-precision ?oating-point number formats needed for guaranteeing the desired accuracy of ?oating-point applications. In this work, we implement the ?rst Matlab-based framework for performing rounding error analysis and numerical range evaluation of arbitrary ?oating-point algorithms using the AA-based error model. Our framework enables users to best reuse their own existing Matlab code to e?ectively conduct rounding error analysis tasks and run bit-true custom-precision computations of ?oating-point algorithms in Matlab for veri?cation. We apply the AA-based error analysis technique and our Matlab-based framework to the ?oating-point rounding error evaluation and optimal uniform bit width allocation of two signal and speech processing applications: i) the ?oating-point dot-product and ii) the iterative Levinson-Durbin algorithm for linear prediction and autoregressive modeling. For the ?oating-point dot-product, it is shown that the AA-based error model can provide tighter rounding error bounds compared to existing error analysis techniques. This corresponds to the overestimation of up to 2 mantissa bits compared to those estimated by running extensive simulations. For the iterative Levinson-Durbin algorithm, the AA-based error analysis technique can model accurately the rounding errors of the coe?cients when using a restricted range for the input parameters. When using a general range for the input parameters, the AA-based error analysis technique can give a qualitative estimate for the error bound of the coe?cients.
