## Reconstruction and clustering with graph optimization and priors on gene networks and images (2017)

Collective analysis of multiple high-throughput gene expression datasets

Modern technologies have resulted in the production of numerous high-throughput biological datasets. However, the pace of development of capable computational methods does not cope with the pace of generation of new high-throughput datasets. Amongst the most popular biological high-throughput datasets are gene expression datasets (e.g. microarray datasets). This work targets this aspect by proposing a suite of computational methods which can analyse multiple gene expression datasets collectively. The focal method in this suite is the unification of clustering results from multiple datasets using external specifications (UNCLES). This method applies clustering to multiple heterogeneous datasets which measure the expression of the same set of genes separately and then combines the resulting partitions in accordance to one of two types of external specifications; type A identifies the subsets of genes that are consistently co-expressed in all of the given datasets while type ...

Abu-Jamous, Basel — Brunel University London

This dissertation deals with the distributed processing techniques for parameter estimation and efficient data-gathering in wireless communication and sensor networks. The estimation problem consists in inferring a set of parameters from temporal and spatial noisy observations collected by different nodes that monitor an area or field. The objective is to derive an estimate that is as accurate as the one that would be obtained if each node had access to the information across the entire network. With the aim of enabling an energy aware and low-complexity distributed implementation of the estimation task, several useful optimization techniques that generally yield linear estimators were derived in the literature. Up to now, most of the works considered that the nodes are interested in estimating the same vector of global parameters. This scenario can be viewed as a special case of a more general ...

Bogdanovic, Nikola — University of Patras

Acoustic sensor network geometry calibration and applications

In the modern world, we are increasingly surrounded by computation devices with communication links and one or more microphones. Such devices are, for example, smartphones, tablets, laptops or hearing aids. These devices can work together as nodes in an acoustic sensor network (ASN). Such networks are a growing platform that opens the possibility for many practical applications. ASN based speech enhancement, source localization, and event detection can be applied for teleconferencing, camera control, automation, or assisted living. For this kind of applications, the awareness of auditory objects and their spatial positioning are key properties. In order to provide these two kinds of information, novel methods have been developed in this thesis. Information on the type of auditory objects is provided by a novel real-time sound classification method. Information on the position of human speakers is provided by a novel localization ...

Plinge, Axel — TU Dortmund University

Accelerating Monte Carlo methods for Bayesian inference in dynamical models

Making decisions and predictions from noisy observations are two important and challenging problems in many areas of society. Some examples of applications are recommendation systems for online shopping and streaming services, connecting genes with certain diseases and modelling climate change. In this thesis, we make use of Bayesian statistics to construct probabilistic models given prior information and historical data, which can be used for decision support and predictions. The main obstacle with this approach is that it often results in mathematical problems lacking analytical solutions. To cope with this, we make use of statistical simulation algorithms known as Monte Carlo methods to approximate the intractable solution. These methods enjoy well-understood statistical properties but are often computational prohibitive to employ. The main contribution of this thesis is the exploration of different strategies for accelerating inference methods based on sequential Monte Carlo ...

Dahlin, Johan — Linköping University

Parametric and non-parametric approaches for multisensor data fusion

Multisensor data fusion technology combines data and information from multiple sensors to achieve improved accuracies and better inference about the environment than could be achieved by the use of a single sensor alone. In this dissertation, we propose parametric and nonparametric multisensor data fusion algorithms with a broad range of applications. Image registration is a vital first step in fusing sensor data. Among the wide range of registration techniques that have been developed for various applications, mutual information based registration algorithms have been accepted as one of the most accurate and robust methods. Inspired by the mutual information based approaches, we propose to use the joint RÂ´enyi entropy as the dissimilarity metric between images. Since the RÂ´enyi entropy of an image can be estimated with the length of the minimum spanning tree over the corresponding graph, the proposed information-theoretic registration ...

Ma, Bing — University of Michigan

Signal processing algorithms for wireless acoustic sensor networks

Recent academic developments have initiated a paradigm shift in the way spatial sensor data can be acquired. Traditional localized and regularly arranged sensor arrays are replaced by sensor nodes that are randomly distributed over the entire spatial field, and which communicate with each other or with a master node through wireless communication links. Together, these nodes form a so-called ‘wireless sensor network’ (WSN). Each node of a WSN has a local sensor array and a signal processing unit to perform computations on the acquired data. The advantage of WSNs compared to traditional (wired) sensor arrays, is that many more sensors can be used that physically cover the full spatial field, which typically yields more variety (and thus more information) in the signals. It is likely that future data acquisition, control and physical monitoring, will heavily rely on this type of ...

Bertrand, Alexander — Katholieke Universiteit Leuven

Variational Sparse Bayesian Learning: Centralized and Distributed Processing

In this thesis we investigate centralized and distributed variants of sparse Bayesian learning (SBL), an effective probabilistic regression method used in machine learning. Since inference in an SBL model is not tractable in closed form, approximations are needed. We focus on the variational Bayesian approximation, as opposed to others used in the literature, for three reasons: First, it is a flexible general framework for approximate Bayesian inference that estimates probability densities including point estimates as a special case. Second, it has guaranteed convergence properties. And third, it is a deterministic approximation concept that is even applicable for high dimensional problems where non-deterministic sampling methods may be prohibitive. We resolve some inconsistencies in the literature involved in other SBL approximation techniques with regard to a proper Bayesian treatment and the incorporation of a very desired property, namely scale invariance. More specifically, ...

Buchgraber, Thomas — Graz University of Technology

Distributed Signal Processing Algorithms for Wireless Networks

Distributed signal processing algorithms have become a key approach for statistical inference in wireless networks and applications such as wireless sensor networks and smart grids. It is well known that distributed processing techniques deal with the extraction of information from data collected at nodes that are distributed over a geographic area. In this context, for each specific node, a set of neighbor nodes collect their local information and transmit the estimates to a specific node. Then, each specific node combines the collected information together with its local estimate to generate an improved estimate. In this thesis, novel distributed cooperative algorithms for inference in ad hoc, wireless sensor networks and smart grids are investigated. Low-complexity and effective algorithms to perform statistical inference in a distributed way are devised. A number of innovative approaches for dealing with node failures, compression of data ...

Xu, Songcen — University of York

Decentralized Estimation Under Communication Constraints

In this thesis, we consider the problem of decentralized estimation under communication constraints in the context of Collaborative Signal and Information Processing. Motivated by sensor network applications, a high volume of data collected at distinct locations and possibly in diverse modalities together with the spatially distributed nature and the resource limitations of the underlying system are of concern. Designing processing schemes which match the constraints imposed by the system while providing a reasonable accuracy has been a major challenge in which we are particularly interested in the tradeoff between the estimation performance and the utilization of communications subject to energy and bandwidth constraints. One remarkable approach for decentralized inference in sensor networks is to exploit graphical models together with message passing algorithms. In this framework, after the so-called information graph of the problem is constructed, it is mapped onto the ...

Uney, Murat — Middle East Technical University

Bayesian Compressed Sensing using Alpha-Stable Distributions

During the last decades, information is being gathered and processed at an explosive rate. This fact gives rise to a very important issue, that is, how to effectively and precisely describe the information content of a given source signal or an ensemble of source signals, such that it can be stored, processed or transmitted by taking into consideration the limitations and capabilities of the several digital devices. One of the fundamental principles of signal processing for decades is the Nyquist-Shannon sampling theorem, which states that the minimum number of samples needed to reconstruct a signal without error is dictated by its bandwidth. However, there are many cases in our everyday life in which sampling at the Nyquist rate results in too many data and thus, demanding an increased processing power, as well as storage requirements. A mathematical theory that emerged ...

Tzagkarakis, George — University of Crete

In recent years, advances in signal processing have led the wireless sensor networks to be capable of mobility. The signal processing in a wireless sensor network differs from that of a traditional wireless network mainly in two important aspects. Unlike traditional wireless networks, in a sensor network the signal processing is performed in a fully distributed manner as the sensor measurements in a distributed fashion across the network collected. Additionally, due to the limited onboard resource of a sensor network it is essential to develop energy and bandwidth efficient signal processing algorithms. The thesis is devoted to discuss the state of the arte of algorithms commonly known as tracking algorithms. Although tracking algorithms have only been attracting research and development attention recently, already a wide literature and great variety of proposed approaches regarding the topic exist. The dissertation focus on ...

Arienzo, Loredana — University of Salerno

Wireless Sensor Networks (WSNs) aim for accurate data gathering and representation of one or multiple physical variables from the environment, by means of sensor reading and wireless data packets transmission to a Data Fusion Center (DFC). There is no comprehensive common set of requirements for all WSN, as they are application dependent. Moreover, due to specific node capabilities or energy consumption constraints several tradeoffs have to be considered during the design, and particularly, the price of the sensor nodes is a determining factor. The distinction between small and large scale WSNs does not only refers to the quantity of sensor nodes, but also establishes the main design challenges in each case. For example, the node organization is a key issue in large scale WSNs, where many inexpensive nodes have to properly work in a coordinated manner. Regarding the amount of ...

Chidean, Mihaela I. — Rey Juan Carlos University

Segmentation par modèle déformable surfacique localement régularisé par spline

Image segmentation through deformable models is a method that localizes object boundaries. When difficult segmentation context are proposed because of noise or a lack of information, the use of prior knowledge in the deformation process increases segmentation accuracy. Medical imaging is often concerned by these context. Moreover, medical applications deal with large amounts of data. Then it is mandatory to use a robust and fast processing. This question lead us to a local regularisation of the deformable model. Highly based on the active contour framework, also known as \emph{snake}, we propose a new regularization scheme. This is done by filtering the displacements at each iteration. The filter is based on a smoothing spline kernel whose aim was to approximate a set of points rather than interpolating it. We point out the consistency of the regularization parameter in such a method. ...

Velut, Jerome — INSA-Lyon / CREATIS-LRMN

Robust Methods for Sensing and Reconstructing Sparse Signals

Compressed sensing (CS) is a recently introduced signal acquisition framework that goes against the traditional Nyquist sampling paradigm. CS demonstrates that a sparse, or compressible, signal can be acquired using a low rate acquisition process. Since noise is always present in practical data acquisition systems, sensing and reconstruction methods are developed assuming a Gaussian (light-tailed) model for the corrupting noise. However, when the underlying signal and/or the measurements are corrupted by impulsive noise, commonly employed linear sampling operators, coupled with Gaussian-derived reconstruction algorithms, fail to recover a close approximation of the signal. This dissertation develops robust sampling and reconstruction methods for sparse signals in the presence of impulsive noise. To achieve this objective, we make use of robust statistics theory to develop appropriate methods addressing the problem of impulsive noise in CS systems. We develop a generalized Cauchy distribution (GCD) ...

Carrillo, Rafael — University of Delaware

In Wireless Sensor Networks (WSN), the ability of sensor nodes to know its position is an enabler for a wide variety of applications for monitoring, control, and automation. Often, sensor data is meaningful only if its position can be determined. Many WSN are deployed indoors or in areas where Global Navigation Satellite System (GNSS) signal coverage is not available, and thus GNSS positioning cannot be guaranteed. In these scenarios, WSN may be relied upon to achieve a satisfactory degree of positioning accuracy. Typically, batteries power sensor nodes in WSN. These batteries are costly to replace. Therefore, power consumption is an important aspect, being performance and lifetime ofWSN strongly relying on the ability to reduce it. It is crucial to design effective strategies to maximize battery lifetime. Optimization of power consumption can be made at different layers. For example, at the ...

Moragrega, Ana — Universitat Politecnica de Catalunya

The current layout is optimized for **mobile
phones**. Page previews, thumbnails, and full abstracts
will remain hidden until the browser window grows in width.

The current layout is optimized for **tablet
devices**. Page previews and some thumbnails will remain
hidden until the browser window grows in width.