Variational Sparse Bayesian Learning: Centralized and Distributed Processing (2013)
Wireless Network Localization via Cooperation
This dissertation details two classes of cooperative localization methods for wireless networks in mixed line-of-sight and non-line-of-sight (LOS/NLOS) environments. The classes of methods depend on the amount of prior knowledge available. The methods used for both classes are based on the assumptions in practical localization environments that neither NLOS identification nor experimental campaigns are affordable. Two major contributions are, first, in methods that provide satisfactory localization accuracy whilst relaxing the requirement on statistical knowledge about the measurement model. Second, in methods that provide significantly improved localization performance without the requirement of good initialization. In the first half of the dissertation, cooperative localization using received signal strength (RSS) measurements in homogeneous mixed LOS/NLOS environments is considered for the case where the key model parameter, the path loss exponent, is unknown. The approach taken is to model the positions and the path ...
Jin, Di — Signal Processing Group, Technische Universität Darmstadt
Decentralized Estimation Under Communication Constraints
In this thesis, we consider the problem of decentralized estimation under communication constraints in the context of Collaborative Signal and Information Processing. Motivated by sensor network applications, a high volume of data collected at distinct locations and possibly in diverse modalities together with the spatially distributed nature and the resource limitations of the underlying system are of concern. Designing processing schemes which match the constraints imposed by the system while providing a reasonable accuracy has been a major challenge in which we are particularly interested in the tradeoff between the estimation performance and the utilization of communications subject to energy and bandwidth constraints. One remarkable approach for decentralized inference in sensor networks is to exploit graphical models together with message passing algorithms. In this framework, after the so-called information graph of the problem is constructed, it is mapped onto the ...
Uney, Murat — Middle East Technical University
Sparse Sensing for Statistical Inference: Theory, Algorithms, and Applications
In today's society, we are flooded with massive volumes of data in the order of a billion gigabytes on a daily basis from pervasive sensors. It is becoming increasingly challenging to locally store and transport the acquired data to a central location for signal/data processing (i.e., for inference). To alleviate these problems, it is evident that there is an urgent need to significantly reduce the sensing cost (i.e., the number of expensive sensors) as well as the related memory and bandwidth requirements by developing unconventional sensing mechanisms to extract as much information as possible yet collecting fewer data. The first aim of this thesis is to develop theory and algorithms for data reduction. We develop a data reduction tool called sparse sensing, which consists of a deterministic and structured sensing function (guided by a sparse vector) that is optimally designed ...
Chepuri, Sundeep Prabhakar — Delft University of Technology
Large-Scale Light Field Capture and Reconstruction
This thesis discusses approaches and techniques to convert Sparsely-Sampled Light Fields (SSLFs) into Densely-Sampled Light Fields (DSLFs), which can be used for visualization on 3DTV and Virtual Reality (VR) devices. Exemplarily, a movable 1D large-scale light field acquisition system for capturing SSLFs in real-world environments is evaluated. This system consists of 24 sparsely placed RGB cameras and two Kinect V2 sensors. The real-world SSLF data captured with this setup can be leveraged to reconstruct real-world DSLFs. To this end, three challenging problems require to be solved for this system: (i) how to estimate the rigid transformation from the coordinate system of a Kinect V2 to the coordinate system of an RGB camera; (ii) how to register the two Kinect V2 sensors with a large displacement; (iii) how to reconstruct a DSLF from a SSLF with moderate and large disparity ranges. ...
Gao, Yuan — Department of Computer Science, Kiel University
Signal Quantization and Approximation Algorithms for Federated Learning
Distributed signal or information processing using Internet of Things (IoT), facilitates real-time monitoring of signals, for example, environmental pollutants, health indicators, and electric energy consumption in a smart city. Despite the promising capabilities of IoTs, these distributed deployments often face the challenge of data privacy and communication rate constraints. In traditional machine learning, training data is moved to a data center, which requires massive data movement from distributed IoT devices to a third-party location, thus raising concerns over privacy and inefficient use of communication resources. Moreover, the growing network size, model size, and data volume combined lead to unusual complexity in the design of optimization algorithms beyond the compute capability of a single device. This necessitates novel system architectures to ensure stable and secure operations of such networks. Federated learning (FL) architecture, a novel distributed learning paradigm introduced by McMahan ...
A, Vijay — Indian Institute of Technology Bombay
Bayesian Approaches in Image Source Seperation
In this thesis, a general solution to the component separation problem in images is introduced. Unlike most existing works, the spatial dependencies of images are modelled in the separation process with the use of Markov random fields (MRFs). In the MRFs model, Cauchy density is used for the gradient images. We provide a general Bayesian framework for the estimation of the parameters of this model. Due to the intractability of the problem we resort to numerical solutions for the joint maximization of the a posteriori distribution of the sources, the mixing matrix and the noise variances. For numerical solution, four different methods are proposed. In first method, the difficulty of working analytically with general Gibbs distributions of MRF is overcome by using an approximate density. In this approach, the Gibbs distribution is modelled by the product of directional Gaussians. The ...
Kayabol, Koray — Istanbul University
Bayesian Compressed Sensing using Alpha-Stable Distributions
During the last decades, information is being gathered and processed at an explosive rate. This fact gives rise to a very important issue, that is, how to effectively and precisely describe the information content of a given source signal or an ensemble of source signals, such that it can be stored, processed or transmitted by taking into consideration the limitations and capabilities of the several digital devices. One of the fundamental principles of signal processing for decades is the Nyquist-Shannon sampling theorem, which states that the minimum number of samples needed to reconstruct a signal without error is dictated by its bandwidth. However, there are many cases in our everyday life in which sampling at the Nyquist rate results in too many data and thus, demanding an increased processing power, as well as storage requirements. A mathematical theory that emerged ...
Tzagkarakis, George — University of Crete
Bayesian methods for sparse and low-rank matrix problems
Many scientific and engineering problems require us to process measurements and data in order to extract information. Since we base decisions on information, it is important to design accurate and efficient processing algorithms. This is often done by modeling the signal of interest and the noise in the problem. One type of modeling is Compressed Sensing, where the signal has a sparse or low-rank representation. In this thesis we study different approaches to designing algorithms for sparse and low-rank problems. Greedy methods are fast methods for sparse problems which iteratively detects and estimates the non-zero components. By modeling the detection problem as an array processing problem and a Bayesian filtering problem, we improve the detection accuracy. Bayesian methods approximate the sparsity by probability distributions which are iteratively modified. We show one approach to making the Bayesian method the Relevance Vector ...
Sundin, Martin — Department of Signal Processing, Royal Institute of Technology KTH
Accelerating Monte Carlo methods for Bayesian inference in dynamical models
Making decisions and predictions from noisy observations are two important and challenging problems in many areas of society. Some examples of applications are recommendation systems for online shopping and streaming services, connecting genes with certain diseases and modelling climate change. In this thesis, we make use of Bayesian statistics to construct probabilistic models given prior information and historical data, which can be used for decision support and predictions. The main obstacle with this approach is that it often results in mathematical problems lacking analytical solutions. To cope with this, we make use of statistical simulation algorithms known as Monte Carlo methods to approximate the intractable solution. These methods enjoy well-understood statistical properties but are often computational prohibitive to employ. The main contribution of this thesis is the exploration of different strategies for accelerating inference methods based on sequential Monte Carlo ...
Dahlin, Johan — Linköping University
Sensing physical fields: Inverse problems for the diffusion equation and beyond
Due to significant advances made over the last few decades in the areas of (wireless) networking, communications and microprocessor fabrication, the use of sensor networks to observe physical phenomena is rapidly becoming commonplace. Over this period, many aspects of sensor networks have been explored, yet a thorough understanding of how to analyse and process the vast amounts of sensor data collected remains an open area of research. This work, therefore, aims to provide theoretical, as well as practical, advances this area. In particular, we consider the problem of inferring certain underlying properties of the monitored phenomena, from our sensor measurements. Within mathematics, this is commonly formulated as an inverse problem; whereas in signal processing, it appears as a (multidimensional) sampling and reconstruction problem. Indeed it is well known that inverse problems are notoriously ill-posed and very demanding to solve; meanwhile ...
Murray-Bruce, John — Imperial College London
Low Complexity Image Recognition Algorithms for Handheld Devices
Content Based Image Retrieval (CBIR) has gained a lot of interest over the last two decades. The need to search and retrieve images from databases, based on information (“features”) extracted from the image itself, is becoming increasingly important. CBIR can be useful for handheld image recognition devices in which the image to be recognized is acquired with a camera, and thus there is no additional metadata associated to it. However, most CBIR systems require large computations, preventing their use in handheld devices. In this PhD work, we have developed low-complexity algorithms for content based image retrieval in handheld devices for camera acquired images. Two novel algorithms, ‘Color Density Circular Crop’ (CDCC) and ‘DCT-Phase Match’ (DCTPM), to perform image retrieval along with a two-stage image retrieval algorithm that combines CDCC and DCTPM, to achieve the low complexity required in handheld devices ...
Ayyalasomayajula, Pradyumna — EPFL
Bayesian data fusion for distributed learning
This dissertation explores the intersection of data fusion, federated learning, and Bayesian methods, with a focus on their applications in indoor localization, GNSS, and image processing. Data fusion involves integrating data and knowledge from multiple sources. It becomes essential when data is only available in a distributed fashion or when different sensors are used to infer a quantity of interest. Data fusion typically includes raw data fusion, feature fusion, and decision fusion. In this thesis, we will concentrate on feature fusion. Distributed data fusion involves merging sensor data from different sources to estimate an unknown process. Bayesian framework is often used because it can provide an optimal and explainable feature by preserving the full distribution of the unknown given the data, called posterior, over the estimated process at each agent. This allows for easy and recursive merging of sensor data ...
Peng Wu — Northeastern University
Short-length Low-density Parity-check Codes: Construction and Decoding Algorithms
Error control coding is an essential part of modern communications systems. LDPC codes have been demonstrated to offer performance near the fundamental limits of channels corrupted by random noise. Optimal maximum likelihood decoding of LDPC codes is too complex to be practically useful even at short block lengths and so a graph-based message passing decoder known as the belief propagation algorithm is used instead. In fact, on graphs without closed paths known as cycles the iterative message passing decoding is known to be optimal and may converge in a single iteration, although identifying the message update schedule which allows single-iteration convergence is not trivial. At finite block lengths graphs without cycles have poor minimum distance properties and perform poorly even under optimal decoding. LDPC codes with large block length have been demonstrated to offer performance close to that predicted for ...
Healy, Cornelius Thomas — University of York
The solution to many image restoration and reconstruction problems is often defined as the minimizer of a penalized criterion that accounts simultaneously for the data and the prior. This thesis deals more specifically with the minimization of edge-preserving penalized criteria. We focus on algorithms for large-scale problems. The minimization of penalized criteria can be addressed using a half-quadratic approach (HQ). Converging HQ algorithms have been proposed. However, their numerical cost is generally too high for large-scale problems. An alternative is to implement inexact HQ algorithms. Nonlinear conjugate gradient algorithms can also be considered using scalar HQ algorithms for the line search (NLCG+HQ1D). Some issues on the convergence of the aforementioned algorithms remained open until now. In this thesis we : - Prove the convergence of inexact HQ algorithms and NLCG+HQ1D. - Point out strong links between HQ algorithms and NLCG+HQ1D. ...
Labat, Christian — IRCCyN, Nantes, France
Robust Wireless Localization in Harsh Mixed Line-of-Sight/Non-Line-of-Sight Environments
This PhD thesis considers the problem of locating some target nodes in different wireless infrastructures such as wireless cellular radio networks and wireless sensor networks. To be as realistic as possible, mixed line-of-sight and non-line-of-sight (LOS/NLOS) localization environment is introduced. Both the conventional non-cooperative localization and the new emerging cooperative localization have been studied thoroughly. Owing to the random nature of the measurements, probabilistic methods are more advanced as compared to the old-fashioned geometric methods. The gist behind the probabilistic methods is to infer the unknown positions of the target nodes in an estimation process, given a set of noisy position related measurements, a probabilistic measurement model, and a few known reference positions. In contrast to the majority of the existing methods, harsh but practical constraints are taken into account: neither offline calibration nor non-line-of-sight state identification is equipped in ...
Yin, Feng — Technische Universität Darmstadt
The current layout is optimized for mobile phones. Page previews, thumbnails, and full abstracts will remain hidden until the browser window grows in width.
The current layout is optimized for tablet devices. Page previews and some thumbnails will remain hidden until the browser window grows in width.